See 
<https://builds.apache.org/job/beam_PostCommit_XVR_Flink/1847/display/redirect?page=changes>

Changes:

[crites] Changes TestStreamTranscriptTest to only emit two elements so that its


------------------------------------------
[...truncated 865.99 KB...]
      string_value: "container"
    }
  }
  fields {
    key: "beam:option:sdk_worker_parallelism:v1"
    value {
      string_value: "1"
    }
  }
  fields {
    key: "beam:option:shutdown_sources_on_final_watermark:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:spark_master_url:v1"
    value {
      string_value: "local[4]"
    }
  }
  fields {
    key: "beam:option:spark_submit_uber_jar:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:streaming:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:style:v1"
    value {
      string_value: "scrambled"
    }
  }
  fields {
    key: "beam:option:type_check_strictness:v1"
    value {
      string_value: "DEFAULT_TO_ANY"
    }
  }
  fields {
    key: "beam:option:update:v1"
    value {
      bool_value: false
    }
  }
  fields {
    key: "beam:option:use_grpc_for_gcs:v1"
    value {
      bool_value: false
    }
  }
}
job_name: "job"

apache_beam.runners.portability.portable_runner: INFO: Job state changed to 
STOPPED
apache_beam.runners.portability.portable_runner: INFO: Job state changed to 
STARTING
apache_beam.runners.portability.portable_runner: INFO: Job state changed to 
RUNNING
root: DEBUG: org.apache.flink.runtime.client.JobExecutionException: Job 
execution failed.
        at 
org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:146)
        at 
org.apache.flink.runtime.minicluster.MiniCluster.executeJobBlocking(MiniCluster.java:626)
        at 
org.apache.flink.client.LocalExecutor.executePlan(LocalExecutor.java:222)
        at 
org.apache.flink.api.java.LocalEnvironment.execute(LocalEnvironment.java:91)
        at 
org.apache.beam.runners.flink.FlinkBatchPortablePipelineTranslator$BatchTranslationContext.execute(FlinkBatchPortablePipelineTranslator.java:194)
        at 
org.apache.beam.runners.flink.FlinkPipelineRunner.runPipelineWithTranslator(FlinkPipelineRunner.java:116)
        at 
org.apache.beam.runners.flink.FlinkPipelineRunner.run(FlinkPipelineRunner.java:84)
        at 
org.apache.beam.runners.fnexecution.jobsubmission.JobInvocation.runPipeline(JobInvocation.java:83)
        at 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
        at 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
        at 
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
Caused by: java.util.concurrent.ExecutionException: java.lang.RuntimeException: 
Error received from SDK harness for instruction 19: 
org.apache.beam.sdk.util.UserCodeException: java.lang.NoClassDefFoundError: 
org/apache/hadoop/conf/Configuration
        at 
org.apache.beam.sdk.util.UserCodeException.wrap(UserCodeException.java:36)
        at 
org.apache.beam.sdk.io.WriteFiles$WriteUnshardedTempFilesFn$DoFnInvoker.invokeProcessElement(Unknown
 Source)
        at 
org.apache.beam.fn.harness.FnApiDoFnRunner.processElementForParDo(FnApiDoFnRunner.java:600)
        at 
org.apache.beam.fn.harness.data.PCollectionConsumerRegistry$MetricTrackingFnDataReceiver.accept(PCollectionConsumerRegistry.java:216)
        at 
org.apache.beam.fn.harness.data.PCollectionConsumerRegistry$MetricTrackingFnDataReceiver.accept(PCollectionConsumerRegistry.java:179)
        at 
org.apache.beam.fn.harness.BeamFnDataReadRunner.forwardElementToConsumer(BeamFnDataReadRunner.java:175)
        at 
org.apache.beam.fn.harness.data.QueueingBeamFnDataClient.drainAndBlock(QueueingBeamFnDataClient.java:106)
        at 
org.apache.beam.fn.harness.control.ProcessBundleHandler.processBundle(ProcessBundleHandler.java:285)
        at 
org.apache.beam.fn.harness.control.BeamFnControlClient.delegateOnInstructionRequestType(BeamFnControlClient.java:160)
        at 
org.apache.beam.fn.harness.control.BeamFnControlClient.lambda$processInstructionRequests$0(BeamFnControlClient.java:144)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.NoClassDefFoundError: org/apache/hadoop/conf/Configuration
        at 
org.apache.parquet.hadoop.ParquetWriter$Builder.<init>(ParquetWriter.java:343)
        at 
org.apache.parquet.avro.AvroParquetWriter$Builder.<init>(AvroParquetWriter.java:162)
        at 
org.apache.parquet.avro.AvroParquetWriter$Builder.<init>(AvroParquetWriter.java:153)
        at 
org.apache.parquet.avro.AvroParquetWriter.builder(AvroParquetWriter.java:43)
        at 
org.apache.beam.sdk.io.parquet.ParquetIO$Sink.open(ParquetIO.java:354)
        at 
org.apache.beam.sdk.io.FileIO$Write$ViaFileBasedSink$1$1.prepareWrite(FileIO.java:1390)
        at 
org.apache.beam.sdk.io.FileBasedSink$Writer.open(FileBasedSink.java:952)
        at 
org.apache.beam.sdk.io.WriteFiles$WriteUnshardedTempFilesFn.processElement(WriteFiles.java:534)

        at 
java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
        at 
java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
        at org.apache.beam.sdk.util.MoreFutures.get(MoreFutures.java:57)
        at 
org.apache.beam.runners.fnexecution.control.SdkHarnessClient$BundleProcessor$ActiveBundle.close(SdkHarnessClient.java:345)
        at 
org.apache.beam.runners.fnexecution.control.DefaultJobBundleFactory$SimpleStageBundleFactory$1.close(DefaultJobBundleFactory.java:402)
        at 
org.apache.beam.runners.flink.translation.functions.FlinkExecutableStageFunction.$closeResource(FlinkExecutableStageFunction.java:204)
        at 
org.apache.beam.runners.flink.translation.functions.FlinkExecutableStageFunction.mapPartition(FlinkExecutableStageFunction.java:204)
        at 
org.apache.flink.runtime.operators.MapPartitionDriver.run(MapPartitionDriver.java:103)
        at org.apache.flink.runtime.operators.BatchTask.run(BatchTask.java:504)
        at 
org.apache.flink.runtime.operators.BatchTask.invoke(BatchTask.java:369)
        at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:705)
        at org.apache.flink.runtime.taskmanager.Task.run(Task.java:530)
        ... 1 more
Caused by: java.lang.RuntimeException: Error received from SDK harness for 
instruction 19: org.apache.beam.sdk.util.UserCodeException: 
java.lang.NoClassDefFoundError: org/apache/hadoop/conf/Configuration
        at 
org.apache.beam.sdk.util.UserCodeException.wrap(UserCodeException.java:36)
        at 
org.apache.beam.sdk.io.WriteFiles$WriteUnshardedTempFilesFn$DoFnInvoker.invokeProcessElement(Unknown
 Source)
        at 
org.apache.beam.fn.harness.FnApiDoFnRunner.processElementForParDo(FnApiDoFnRunner.java:600)
        at 
org.apache.beam.fn.harness.data.PCollectionConsumerRegistry$MetricTrackingFnDataReceiver.accept(PCollectionConsumerRegistry.java:216)
        at 
org.apache.beam.fn.harness.data.PCollectionConsumerRegistry$MetricTrackingFnDataReceiver.accept(PCollectionConsumerRegistry.java:179)
        at 
org.apache.beam.fn.harness.BeamFnDataReadRunner.forwardElementToConsumer(BeamFnDataReadRunner.java:175)
        at 
org.apache.beam.fn.harness.data.QueueingBeamFnDataClient.drainAndBlock(QueueingBeamFnDataClient.java:106)
        at 
org.apache.beam.fn.harness.control.ProcessBundleHandler.processBundle(ProcessBundleHandler.java:285)
        at 
org.apache.beam.fn.harness.control.BeamFnControlClient.delegateOnInstructionRequestType(BeamFnControlClient.java:160)
        at 
org.apache.beam.fn.harness.control.BeamFnControlClient.lambda$processInstructionRequests$0(BeamFnControlClient.java:144)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.NoClassDefFoundError: org/apache/hadoop/conf/Configuration
        at 
org.apache.parquet.hadoop.ParquetWriter$Builder.<init>(ParquetWriter.java:343)
        at 
org.apache.parquet.avro.AvroParquetWriter$Builder.<init>(AvroParquetWriter.java:162)
        at 
org.apache.parquet.avro.AvroParquetWriter$Builder.<init>(AvroParquetWriter.java:153)
        at 
org.apache.parquet.avro.AvroParquetWriter.builder(AvroParquetWriter.java:43)
        at 
org.apache.beam.sdk.io.parquet.ParquetIO$Sink.open(ParquetIO.java:354)
        at 
org.apache.beam.sdk.io.FileIO$Write$ViaFileBasedSink$1$1.prepareWrite(FileIO.java:1390)
        at 
org.apache.beam.sdk.io.FileBasedSink$Writer.open(FileBasedSink.java:952)
        at 
org.apache.beam.sdk.io.WriteFiles$WriteUnshardedTempFilesFn.processElement(WriteFiles.java:534)

        at 
org.apache.beam.runners.fnexecution.control.FnApiControlClient$ResponseStreamObserver.onNext(FnApiControlClient.java:160)
        at 
org.apache.beam.runners.fnexecution.control.FnApiControlClient$ResponseStreamObserver.onNext(FnApiControlClient.java:140)
        at 
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.stub.ServerCalls$StreamingServerCallHandler$StreamingServerCallListener.onMessage(ServerCalls.java:251)
        at 
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.ForwardingServerCallListener.onMessage(ForwardingServerCallListener.java:33)
        at 
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.Contexts$ContextualizedServerCallListener.onMessage(Contexts.java:76)
        at 
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.messagesAvailableInternal(ServerCallImpl.java:309)
        at 
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.messagesAvailable(ServerCallImpl.java:292)
        at 
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1MessagesAvailable.runInContext(ServerImpl.java:782)
        at 
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
        at 
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        ... 1 more

root: ERROR: java.lang.RuntimeException: Error received from SDK harness for 
instruction 19: org.apache.beam.sdk.util.UserCodeException: 
java.lang.NoClassDefFoundError: org/apache/hadoop/conf/Configuration
        at 
org.apache.beam.sdk.util.UserCodeException.wrap(UserCodeException.java:36)
        at 
org.apache.beam.sdk.io.WriteFiles$WriteUnshardedTempFilesFn$DoFnInvoker.invokeProcessElement(Unknown
 Source)
        at 
org.apache.beam.fn.harness.FnApiDoFnRunner.processElementForParDo(FnApiDoFnRunner.java:600)
        at 
org.apache.beam.fn.harness.data.PCollectionConsumerRegistry$MetricTrackingFnDataReceiver.accept(PCollectionConsumerRegistry.java:216)
        at 
org.apache.beam.fn.harness.data.PCollectionConsumerRegistry$MetricTrackingFnDataReceiver.accept(PCollectionConsumerRegistry.java:179)
        at 
org.apache.beam.fn.harness.BeamFnDataReadRunner.forwardElementToConsumer(BeamFnDataReadRunner.java:175)
        at 
org.apache.beam.fn.harness.data.QueueingBeamFnDataClient.drainAndBlock(QueueingBeamFnDataClient.java:106)
        at 
org.apache.beam.fn.harness.control.ProcessBundleHandler.processBundle(ProcessBundleHandler.java:285)
        at 
org.apache.beam.fn.harness.control.BeamFnControlClient.delegateOnInstructionRequestType(BeamFnControlClient.java:160)
        at 
org.apache.beam.fn.harness.control.BeamFnControlClient.lambda$processInstructionRequests$0(BeamFnControlClient.java:144)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.NoClassDefFoundError: org/apache/hadoop/conf/Configuration
        at 
org.apache.parquet.hadoop.ParquetWriter$Builder.<init>(ParquetWriter.java:343)
        at 
org.apache.parquet.avro.AvroParquetWriter$Builder.<init>(AvroParquetWriter.java:162)
        at 
org.apache.parquet.avro.AvroParquetWriter$Builder.<init>(AvroParquetWriter.java:153)
        at 
org.apache.parquet.avro.AvroParquetWriter.builder(AvroParquetWriter.java:43)
        at 
org.apache.beam.sdk.io.parquet.ParquetIO$Sink.open(ParquetIO.java:354)
        at 
org.apache.beam.sdk.io.FileIO$Write$ViaFileBasedSink$1$1.prepareWrite(FileIO.java:1390)
        at 
org.apache.beam.sdk.io.FileBasedSink$Writer.open(FileBasedSink.java:952)
        at 
org.apache.beam.sdk.io.WriteFiles$WriteUnshardedTempFilesFn.processElement(WriteFiles.java:534)

apache_beam.runners.portability.portable_runner: INFO: Job state changed to 
FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: nosetests-xlangValidateRunner.xml
----------------------------------------------------------------------
XML: 
<https://builds.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 10 tests in 211.865s

FAILED (errors=1)

> Task 
> :runners:flink:1.9:job-server:validatesCrossLanguageRunnerPythonUsingJava 
> FAILED

> Task 
> :runners:flink:1.9:job-server:validatesCrossLanguageRunnerPythonUsingPython
setup.py:252: UserWarning: You are using Apache Beam with Python 2. New 
releases of Apache Beam will soon support Python 3 only.
  'You are using Apache Beam with Python 2. '
<https://builds.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/build/gradleenv/1922375555/local/lib/python2.7/site-packages/setuptools/dist.py>:476:
 UserWarning: Normalizing '2.20.0.dev' to '2.20.0.dev0'
  normalized_version,
warning: no files found matching 'README.md'
warning: no files found matching 'NOTICE'
warning: no files found matching 'LICENSE'
<https://builds.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/python/apache_beam/__init__.py>:82:
 UserWarning: You are using Apache Beam with Python 2. New releases of Apache 
Beam will soon support Python 3 only.
  'You are using Apache Beam with Python 2. '
<https://builds.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/python/apache_beam/transforms/trigger_test.py>:541:
 YAMLLoadWarning: calling yaml.load_all() without Loader=... is deprecated, as 
the default Loader is unsafe. Please read https://msg.pyyaml.org/load for full 
details.
  for spec in yaml.load_all(open(transcript_filename)):
test_generate_sequence 
(apache_beam.io.external.generate_sequence_test.XlangGenerateSequenceTest) ... 
ok
test_write (apache_beam.io.external.xlang_parquetio_test.XlangParquetIOTest) 
... ok
test_cogroup_by_key 
(apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest) ... 
ok
test_combine_globally 
(apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest) ... 
ok
test_combine_per_key 
(apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest) ... 
ok
test_flatten 
(apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest) ... 
ok
test_group_by_key 
(apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest) ... 
ok
test_multi_input_output_with_sideinput 
(apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest) ... 
ok
test_partition 
(apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest) ... 
ok
test_prefix 
(apache_beam.transforms.validate_runner_xlang_test.ValidateRunnerXlangTest) ... 
ok

----------------------------------------------------------------------
XML: nosetests-xlangValidateRunner.xml
----------------------------------------------------------------------
XML: 
<https://builds.apache.org/job/beam_PostCommit_XVR_Flink/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 10 tests in 150.408s

OK

> Task :runners:flink:1.9:job-server:validatesCrossLanguageRunnerCleanup
> Task :runners:flink:1.9:job-server:flinkJobServerCleanup

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task 
':runners:flink:1.9:job-server:validatesCrossLanguageRunnerPythonUsingJava'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug 
option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with 
Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See 
https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 26m 40s
113 actionable tasks: 86 executed, 25 from cache, 2 up-to-date

Publishing build scan...
https://gradle.com/s/bgsypo7fuvteu

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to