See 
<https://ci-beam.apache.org/job/beam_LoadTests_Python_Combine_Flink_Batch/382/display/redirect?page=changes>

Changes:

[Kamil Wasilewski] [BEAM-10674] Python CoGBK: Remove unnecessary step. Adjust 
parameters to

[chamikaramj] Include coders embedded in CombinePayload when converting x-lang

[zyichi] Allow Nexmark launcher to publish human-readable events to pubsub.

[Kamil Wasilewski] [BEAM-10674] Python CoGBK: Add streaming job

[douglas.damon] Add Filter lesson to Go SDK katas

[douglas.damon] Reformat code for readability

[yoshiki.obata] [BEAM-9980] Configure Python versions for direct test suite 
tasks via

[irvi.fa] [BEAM-10753] Add Slack link invitation on README

[Luke Cwik] [BEAM-10756] Fix empty pull response to not ack and to not throw

[ettarapp] clarifying unclear comments

[noreply] Update README.md

[noreply] [BEAM-3301] Adding SDF Go Dataflow translation. (#12629)

[kevinsijo] [BEAM-9919] Refining xlang types and wrappers

[noreply] [BEAM-10752] Use TestPubsubSignal in PubsubToBigqueryIT (#12625)

[douglas.damon] Change license info format in task-info.yaml

[douglas.damon] Update stepik course

[Luke Cwik] [BEAM-9979] Support a reset call that happens after the bundle is 
done

[noreply] [BEAM-10200] Add optional experiment to enable heap dump through the …

[kevinsijo] [BEAM-9919] xlang transforms wrapped with namespaces

[kevinsijo] [BEAM-9919] cleaning namespace.go

[kevinsijo] [BEAM-9920] Enabling artifact staging for xlang transforms

[kevinsijo] documentation(beam/sdks/go): added apache license and function 
docstring

[noreply] [BEAM-9547] Implement dataframes top, join, merge. (#12516)

[kevinsijo] fix(beam/sdks/go): correcting changed function name

[noreply] Merge pull request #12581 from [BEAM-10378] Add Azure Blob Storage


------------------------------------------
[...truncated 59.34 KB...]
> Task :sdks:java:io:mongodb:processResources NO-SOURCE
> Task :sdks:java:io:parquet:processResources NO-SOURCE
> Task :sdks:java:extensions:sql:zetasql:processResources NO-SOURCE
> Task 
> :sdks:java:extensions:sql:expansion-service:createCheckerFrameworkManifest
> Task :sdks:java:extensions:protobuf:processResources NO-SOURCE
> Task :sdks:java:extensions:sql:expansion-service:processResources NO-SOURCE
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :runners:flink:1.10:job-server-container:copyLicenses
> Task :runners:flink:1.10:job-server-container:dockerClean UP-TO-DATE
> Task :model:job-management:processResources
> Task :runners:flink:1.10:copySourceOverrides
> Task :model:fn-execution:processResources
> Task :runners:flink:1.10:copyTestResourcesOverrides NO-SOURCE
> Task :runners:flink:1.10:processResources
> Task :sdks:java:core:processResources
> Task :sdks:java:build-tools:compileJava FROM-CACHE
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:build-tools:jar
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromCalciteCore
> Task :sdks:java:extensions:sql:copyFmppTemplatesFromSrc
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :sdks:java:extensions:sql:generateFmppSources
> Task :model:pipeline:jar
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes

> Task :sdks:java:extensions:sql:compileJavacc
Java Compiler Compiler Version 4.0 (Parser Generator)
(type "javacc" with no arguments for help)
Reading from file 
<https://ci-beam.apache.org/job/beam_LoadTests_Python_Combine_Flink_Batch/ws/src/sdks/java/extensions/sql/build/generated/fmpp/javacc/Parser.jj>
 . . .
Note: UNICODE_INPUT option is specified. Please make sure you create the 
parser/lexer using a Reader with the correct character encoding.
Warning: Lookahead adequacy checking not being performed since option LOOKAHEAD 
is more than 1.  Set option FORCE_LA_CHECK to true to force checking.
File "TokenMgrError.java" does not exist.  Will create one.
File "ParseException.java" does not exist.  Will create one.
File "Token.java" does not exist.  Will create one.
File "SimpleCharStream.java" does not exist.  Will create one.
Parser generated with 0 errors and 1 warnings.

> Task :sdks:java:extensions:sql:processResources
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :sdks:java:extensions:join-library:compileJava FROM-CACHE
> Task :sdks:java:extensions:join-library:classes UP-TO-DATE
> Task :runners:local-java:compileJava FROM-CACHE
> Task :runners:local-java:classes UP-TO-DATE
> Task :runners:local-java:jar
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :sdks:java:io:mongodb:compileJava FROM-CACHE
> Task :sdks:java:io:mongodb:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:extractIncludeProto
> Task :sdks:java:extensions:protobuf:generateProto NO-SOURCE
> Task :sdks:java:extensions:join-library:jar
> Task :sdks:java:io:mongodb:jar
> Task :sdks:java:extensions:protobuf:compileJava FROM-CACHE
> Task :sdks:java:extensions:protobuf:classes UP-TO-DATE
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:extensions:protobuf:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :runners:core-construction-java:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:io:hadoop-common:compileJava FROM-CACHE
> Task :sdks:java:io:hadoop-common:classes UP-TO-DATE
> Task :sdks:java:io:hadoop-common:jar
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:io:parquet:compileJava FROM-CACHE
> Task :sdks:java:io:parquet:classes UP-TO-DATE
> Task :sdks:java:io:parquet:jar
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:core:jar
> Task :sdks:java:harness:jar
> Task :sdks:java:harness:shadowJar
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar
> Task :sdks:java:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:expansion-service:classes UP-TO-DATE
> Task :sdks:java:expansion-service:jar
> Task :runners:java-job-service:compileJava FROM-CACHE
> Task :runners:java-job-service:classes UP-TO-DATE
> Task :runners:direct-java:compileJava FROM-CACHE
> Task :runners:direct-java:classes UP-TO-DATE
> Task :runners:java-job-service:jar
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:io:kafka:jar
> Task :runners:flink:1.10:compileJava FROM-CACHE
> Task :runners:flink:1.10:classes
> Task :sdks:java:io:google-cloud-platform:compileJava FROM-CACHE
> Task :sdks:java:io:google-cloud-platform:classes UP-TO-DATE
> Task :runners:flink:1.10:jar
> Task :runners:flink:1.10:job-server:compileJava NO-SOURCE
> Task :runners:flink:1.10:job-server:classes UP-TO-DATE
> Task :sdks:java:io:google-cloud-platform:jar
> Task :runners:direct-java:shadowJar
> Task :sdks:java:extensions:sql:compileJava FROM-CACHE
> Task :sdks:java:extensions:sql:classes
> Task :sdks:java:extensions:sql:jar
> Task :sdks:java:extensions:sql:zetasql:compileJava FROM-CACHE
> Task :sdks:java:extensions:sql:zetasql:classes UP-TO-DATE
> Task :sdks:java:extensions:sql:zetasql:jar
> Task :sdks:java:extensions:sql:expansion-service:compileJava FROM-CACHE
> Task :sdks:java:extensions:sql:expansion-service:classes UP-TO-DATE
> Task :sdks:java:extensions:sql:expansion-service:jar
> Task :sdks:java:extensions:sql:expansion-service:shadowJar
> Task :runners:flink:1.10:job-server:shadowJar
> Task :runners:flink:1.10:job-server-container:copyDockerfileDependencies
> Task :runners:flink:1.10:job-server-container:dockerPrepare
> Task :runners:flink:1.10:job-server-container:docker

Deprecated Gradle features were used in this build, making it incompatible with 
Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See 
https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD SUCCESSFUL in 2m 14s
93 actionable tasks: 64 executed, 28 from cache, 1 up-to-date

Publishing build scan...
https://gradle.com/s/iqnweaoqxwa7i

[beam_LoadTests_Python_Combine_Flink_Batch] $ /bin/bash -xe 
/tmp/jenkins5436420601824329679.sh
+ echo 'Tagging image...'
Tagging image...
[beam_LoadTests_Python_Combine_Flink_Batch] $ /bin/bash -xe 
/tmp/jenkins8287645068458217885.sh
+ docker tag 
gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest 
gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest
[beam_LoadTests_Python_Combine_Flink_Batch] $ /bin/bash -xe 
/tmp/jenkins7410419119090605789.sh
+ echo 'Pushing image...'
Pushing image...
[beam_LoadTests_Python_Combine_Flink_Batch] $ /bin/bash -xe 
/tmp/jenkins7510636751204169908.sh
+ docker push 
gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest
The push refers to repository 
[gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server]
f7f61e9cb72e: Preparing
736199154bb5: Preparing
dc52038509e6: Preparing
9b0b24065d39: Preparing
d9495c5ba4af: Preparing
fd48b7313a1f: Preparing
16e8bdbf703d: Preparing
de9aadc6b492: Preparing
e5df62d9b33a: Preparing
7a9460d53218: Preparing
b2765ac0333a: Preparing
0ced13fcf944: Preparing
e5df62d9b33a: Waiting
7a9460d53218: Waiting
b2765ac0333a: Waiting
0ced13fcf944: Waiting
fd48b7313a1f: Waiting
16e8bdbf703d: Waiting
de9aadc6b492: Waiting
736199154bb5: Pushed
dc52038509e6: Pushed
f7f61e9cb72e: Pushed
fd48b7313a1f: Layer already exists
16e8bdbf703d: Layer already exists
de9aadc6b492: Layer already exists
e5df62d9b33a: Layer already exists
7a9460d53218: Layer already exists
b2765ac0333a: Layer already exists
0ced13fcf944: Layer already exists
d9495c5ba4af: Pushed
9b0b24065d39: Pushed
latest: digest: 
sha256:829681424d63919b1c76f0a82c067dfbbfc3132d56500a2d6a2c73d7c9c4463b size: 
2841
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties content 
JOB_SERVER_IMAGE=gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest
CLUSTER_NAME=beam-loadtests-python-combine-flink-batch-382
DETACHED_MODE=true
HARNESS_IMAGES_TO_PULL=gcr.io/apache-beam-testing/beam_portability/beam_python3.7_sdk:latest
FLINK_NUM_WORKERS=16
FLINK_DOWNLOAD_URL=https://archive.apache.org/dist/flink/flink-1.10.1/flink-1.10.1-bin-scala_2.11.tgz
GCS_BUCKET=gs://beam-flink-cluster
HADOOP_DOWNLOAD_URL=https://repo.maven.apache.org/maven2/org/apache/flink/flink-shaded-hadoop-2-uber/2.8.3-9.0/flink-shaded-hadoop-2-uber-2.8.3-9.0.jar
FLINK_TASKMANAGER_SLOTS=1
ARTIFACTS_DIR=gs://beam-flink-cluster/beam-loadtests-python-combine-flink-batch-382
GCLOUD_ZONE=us-central1-a

[EnvInject] - Variables injected successfully.
[beam_LoadTests_Python_Combine_Flink_Batch] $ /bin/bash -xe 
/tmp/jenkins238033776517759913.sh
+ echo Setting up flink cluster
Setting up flink cluster
[beam_LoadTests_Python_Combine_Flink_Batch] $ /bin/bash -xe 
/tmp/jenkins4746513168846961533.sh
+ cd 
<https://ci-beam.apache.org/job/beam_LoadTests_Python_Combine_Flink_Batch/ws/src/.test-infra/dataproc>
+ ./flink_cluster.sh create
+ GCLOUD_ZONE=us-central1-a
+ DATAPROC_VERSION=1.2
+ MASTER_NAME=beam-loadtests-python-combine-flink-batch-382-m
+ INIT_ACTIONS_FOLDER_NAME=init-actions
+ FLINK_INIT=gs://beam-flink-cluster/init-actions/flink.sh
+ BEAM_INIT=gs://beam-flink-cluster/init-actions/beam.sh
+ DOCKER_INIT=gs://beam-flink-cluster/init-actions/docker.sh
+ FLINK_LOCAL_PORT=8081
+ FLINK_TASKMANAGER_SLOTS=1
+ YARN_APPLICATION_MASTER=
+ create
+ upload_init_actions
+ echo 'Uploading initialization actions to GCS bucket: gs://beam-flink-cluster'
Uploading initialization actions to GCS bucket: gs://beam-flink-cluster
+ gsutil cp -r init-actions/beam.sh init-actions/docker.sh 
init-actions/flink.sh gs://beam-flink-cluster/init-actions
Copying file://init-actions/beam.sh [Content-Type=text/x-sh]...
/ [0 files][    0.0 B/  2.3 KiB]                                                
/ [1 files][  2.3 KiB/  2.3 KiB]                                                
Copying file://init-actions/docker.sh [Content-Type=text/x-sh]...
/ [1 files][  2.3 KiB/  6.0 KiB]                                                
/ [2 files][  6.0 KiB/  6.0 KiB]                                                
-Copying file://init-actions/flink.sh [Content-Type=text/x-sh]...
- [2 files][  6.0 KiB/ 13.7 KiB]                                                
- [3 files][ 13.7 KiB/ 13.7 KiB]                                                
Operation completed over 3 objects/13.7 KiB.                                    
 
+ create_cluster
+ local 
metadata=flink-snapshot-url=https://archive.apache.org/dist/flink/flink-1.10.1/flink-1.10.1-bin-scala_2.11.tgz,
+ metadata+=flink-start-yarn-session=true,
+ metadata+=flink-taskmanager-slots=1,
+ 
metadata+=hadoop-jar-url=https://repo.maven.apache.org/maven2/org/apache/flink/flink-shaded-hadoop-2-uber/2.8.3-9.0/flink-shaded-hadoop-2-uber-2.8.3-9.0.jar
+ [[ -n gcr.io/apache-beam-testing/beam_portability/beam_python3.7_sdk:latest ]]
+ 
metadata+=,beam-sdk-harness-images-to-pull=gcr.io/apache-beam-testing/beam_portability/beam_python3.7_sdk:latest
+ [[ -n 
gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest ]]
+ 
metadata+=,beam-job-server-image=gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest
+ local image_version=1.2
+ echo 'Starting dataproc cluster. Dataproc version: 1.2'
Starting dataproc cluster. Dataproc version: 1.2
+ local num_dataproc_****s=17
+ gcloud dataproc clusters create beam-loadtests-python-combine-flink-batch-382 
--region=global --num-****s=17 --initialization-actions 
gs://beam-flink-cluster/init-actions/docker.sh,gs://beam-flink-cluster/init-actions/beam.sh,gs://beam-flink-cluster/init-actions/flink.sh
 --metadata 
flink-snapshot-url=https://archive.apache.org/dist/flink/flink-1.10.1/flink-1.10.1-bin-scala_2.11.tgz,flink-start-yarn-session=true,flink-taskmanager-slots=1,hadoop-jar-url=https://repo.maven.apache.org/maven2/org/apache/flink/flink-shaded-hadoop-2-uber/2.8.3-9.0/flink-shaded-hadoop-2-uber-2.8.3-9.0.jar,beam-sdk-harness-images-to-pull=gcr.io/apache-beam-testing/beam_portability/beam_python3.7_sdk:latest,beam-job-server-image=gcr.io/apache-beam-testing/beam_portability/beam_flink1.10_job_server:latest,
 --image-version=1.2 --zone=us-central1-a --quiet
Waiting on operation 
[projects/apache-beam-testing/regions/global/operations/761acdd1-f540-3d48-80f0-673e794436fd].
Waiting for cluster creation operation...
WARNING: For PD-Standard without local SSDs, we strongly recommend provisioning 
1TB or larger to ensure consistently high I/O performance. See 
https://cloud.google.com/compute/docs/disks/performance for information on disk 
I/O performance.
................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................done.
ERROR: (gcloud.dataproc.clusters.create) Operation 
[projects/apache-beam-testing/regions/global/operations/761acdd1-f540-3d48-80f0-673e794436fd]
 failed: Initialization action timed out. Failed action 
'gs://beam-flink-cluster/init-actions/flink.sh', see output in: 
gs://dataproc-6c5fbcbb-a2de-406e-9cf7-8c1ce0b6a604-us/google-cloud-dataproc-metainfo/44e72ace-c358-4d0a-9988-6c8a9ccc23c4/beam-loadtests-python-combine-flink-batch-382-w-16/dataproc-initialization-script-2_output.
Build step 'Execute shell' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to