See 
<https://builds.apache.org/job/beam_LoadTests_Python_GBK_Flink_Batch/8/display/redirect?page=changes>

Changes:

[ehudm] [BEAM-6877] Fix trivial_inference for Python 3.x

[ehudm] Address review comments

[je.ik] [BEAM-7543] ReduceByKey.combineBy must accept BinaryFunction<V, V, V>

[ehudm] [BEAM-7579] Use bucket with default key in ITs

[ehudm] Allow more time for IT pipeline to run

[ehudm] Address review comments

[sniemitz] [BEAM-7507] Avoid more String.format and ByteString.toStringUtf8 in 
hot

[owenzhang1990] [BEAM-7467] Do not run gearpump pipeline remotely by default

[je.ik] [BEAM-7543] deprecate old #combineBy(Sums.ofLongs(),

[dcavazos] Add common code infrastructure for element-wise snippets

[alireza4263] [BEAM-7513] Adding RowCount to BigQueryTable.

[aaltay] [BEAM-6955] Use base version component of Beam Python SDK version when

[kedin] Moving to 2.15.0-SNAPSHOT on master branch.

[github] [BEAM-7603] Adding support for ValueProvider in Custom GCS Temp 
Location

[github] Add link to gRPC.io on website portability page

[agoos] [BEAM7597] Fix Typo on website

------------------------------------------
[...truncated 52.74 KB...]
> Task :sdks:java:harness:processResources NO-SOURCE
> Task :sdks:java:core:generateAvroJava NO-SOURCE
> Task :runners:java-fn-execution:processResources NO-SOURCE
> Task :runners:flink:1.7:job-server:processResources NO-SOURCE
> Task :runners:flink:1.7:job-server-container:dockerClean UP-TO-DATE
> Task :sdks:java:io:kafka:processResources NO-SOURCE
> Task :model:job-management:extractProto
> Task :model:fn-execution:extractProto
> Task :runners:flink:1.7:processResources
> Task :sdks:java:core:generateGrammarSource FROM-CACHE
> Task :model:job-management:processResources
> Task :model:fn-execution:processResources
> Task :sdks:java:core:processResources
> Task :model:pipeline:extractIncludeProto
> Task :model:pipeline:extractProto
> Task :model:pipeline:generateProto
> Task :model:pipeline:compileJava FROM-CACHE
> Task :model:pipeline:processResources
> Task :model:pipeline:classes
> Task :model:pipeline:jar
> Task :model:job-management:extractIncludeProto
> Task :model:fn-execution:extractIncludeProto
> Task :model:job-management:generateProto
> Task :sdks:java:build-tools:compileJava
> Task :sdks:java:build-tools:processResources
> Task :sdks:java:build-tools:classes
> Task :sdks:java:build-tools:jar
> Task :model:fn-execution:generateProto
> Task :model:job-management:compileJava FROM-CACHE
> Task :model:job-management:classes
> Task :model:fn-execution:compileJava FROM-CACHE
> Task :model:fn-execution:classes
> Task :model:pipeline:shadowJar
> Task :model:job-management:shadowJar
> Task :model:fn-execution:shadowJar
> Task :sdks:java:core:compileJava FROM-CACHE
> Task :sdks:java:core:classes
> Task :sdks:java:core:shadowJar
> Task :vendor:sdks-java-extensions-protobuf:compileJava FROM-CACHE
> Task :vendor:sdks-java-extensions-protobuf:classes UP-TO-DATE
> Task :runners:core-construction-java:compileJava FROM-CACHE
> Task :runners:core-construction-java:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:compileJava FROM-CACHE
> Task :sdks:java:extensions:google-cloud-platform-core:classes UP-TO-DATE
> Task :sdks:java:io:kafka:compileJava FROM-CACHE
> Task :sdks:java:io:kafka:classes UP-TO-DATE
> Task :sdks:java:extensions:google-cloud-platform-core:jar
> Task :sdks:java:io:kafka:jar
> Task :sdks:java:fn-execution:compileJava FROM-CACHE
> Task :sdks:java:fn-execution:classes UP-TO-DATE
> Task :sdks:java:fn-execution:jar
> Task :runners:core-construction-java:jar
> Task :runners:core-java:compileJava FROM-CACHE
> Task :runners:core-java:classes UP-TO-DATE
> Task :runners:core-java:jar
> Task :sdks:java:harness:compileJava FROM-CACHE
> Task :sdks:java:harness:classes UP-TO-DATE
> Task :sdks:java:harness:jar
> Task :vendor:sdks-java-extensions-protobuf:shadowJar
> Task :sdks:java:harness:shadowJar
> Task :runners:java-fn-execution:compileJava FROM-CACHE
> Task :runners:java-fn-execution:classes UP-TO-DATE
> Task :runners:java-fn-execution:jar

> Task :runners:flink:1.7:compileJava
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Note: Some input files use unchecked or unsafe operations.
Note: Recompile with -Xlint:unchecked for details.

> Task :runners:flink:1.7:classes
> Task :runners:flink:1.7:jar
> Task :runners:flink:1.7:job-server:compileJava NO-SOURCE
> Task :runners:flink:1.7:job-server:classes UP-TO-DATE
> Task :runners:flink:1.7:job-server:shadowJar
> Task :runners:flink:1.7:job-server-container:copyDockerfileDependencies
> Task :runners:flink:1.7:job-server-container:dockerPrepare
> Task :runners:flink:1.7:job-server-container:docker

Deprecated Gradle features were used in this build, making it incompatible with 
Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See 
https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD SUCCESSFUL in 1m 7s
51 actionable tasks: 37 executed, 13 from cache, 1 up-to-date

Publishing build scan...
https://gradle.com/s/xsb3bo4ebqky2

[beam_LoadTests_Python_GBK_Flink_Batch] $ /bin/bash -xe 
/tmp/jenkins8456783824975490087.sh
+ echo 'Tagging Flink Job Server'\''s image...'
Tagging Flink Job Server's image...
[beam_LoadTests_Python_GBK_Flink_Batch] $ /bin/bash -xe 
/tmp/jenkins4762927132897566937.sh
+ docker tag gcr.io/apache-beam-testing/beam_portability/flink-job-server 
gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest
[beam_LoadTests_Python_GBK_Flink_Batch] $ /bin/bash -xe 
/tmp/jenkins5459170222145850682.sh
+ echo 'Pushing Flink Job Server'\''s image...'
Pushing Flink Job Server's image...
[beam_LoadTests_Python_GBK_Flink_Batch] $ /bin/bash -xe 
/tmp/jenkins1748736859846160284.sh
+ docker push 
gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest
The push refers to repository 
[gcr.io/apache-beam-testing/beam_portability/flink-job-server]
0174602e9c9e: Preparing
f17ef791b886: Preparing
7f7647c10076: Preparing
f7d12d471667: Preparing
f350d0146bb3: Preparing
e38df31d449c: Preparing
af5ae4841776: Preparing
b17cc31e431b: Preparing
12cb127eee44: Preparing
604829a174eb: Preparing
fbb641a8b943: Preparing
af5ae4841776: Waiting
b17cc31e431b: Waiting
12cb127eee44: Waiting
e38df31d449c: Waiting
604829a174eb: Waiting
fbb641a8b943: Waiting
f350d0146bb3: Layer already exists
f7d12d471667: Layer already exists
e38df31d449c: Layer already exists
af5ae4841776: Layer already exists
b17cc31e431b: Layer already exists
12cb127eee44: Layer already exists
604829a174eb: Layer already exists
fbb641a8b943: Layer already exists
0174602e9c9e: Pushed
7f7647c10076: Pushed
f17ef791b886: Pushed
latest: digest: 
sha256:f8be02c57445ee1a858995885121d94f3d7501a4d00bc257f682a6f41c9b1fe0 size: 
2632
[EnvInject] - Injecting environment variables from a build step.
[EnvInject] - Injecting as environment variables the properties content 
JOB_SERVER_IMAGE=gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest
CLUSTER_NAME=beam-loadtests-python-gbk-flink-batch-8
DETACHED_MODE=true
HARNESS_IMAGES_TO_PULL=gcr.io/apache-beam-testing/beam_portability/python:latest
FLINK_NUM_WORKERS=5
FLINK_DOWNLOAD_URL=https://archive.apache.org/dist/flink/flink-1.7.0/flink-1.7.0-bin-hadoop28-scala_2.11.tgz
GCS_BUCKET=gs://beam-flink-cluster
FLINK_TASKMANAGER_SLOTS=1
ARTIFACTS_DIR=gs://beam-flink-cluster/beam-loadtests-python-gbk-flink-batch-8
GCLOUD_ZONE=us-central1-a

[EnvInject] - Variables injected successfully.
[beam_LoadTests_Python_GBK_Flink_Batch] $ /bin/bash -xe 
/tmp/jenkins5627198602943210097.sh
+ echo Setting up flink cluster
Setting up flink cluster
[beam_LoadTests_Python_GBK_Flink_Batch] $ /bin/bash -xe 
/tmp/jenkins5764741823248977671.sh
+ cd 
<https://builds.apache.org/job/beam_LoadTests_Python_GBK_Flink_Batch/ws/src/.test-infra/dataproc/>
+ ./create_flink_cluster.sh
+ GCLOUD_ZONE=us-central1-a
+ DATAPROC_VERSION=1.2
+ MASTER_NAME=beam-loadtests-python-gbk-flink-batch-8-m
+ INIT_ACTIONS_FOLDER_NAME=init-actions
+ FLINK_INIT=gs://beam-flink-cluster/init-actions/flink.sh
+ BEAM_INIT=gs://beam-flink-cluster/init-actions/beam.sh
+ DOCKER_INIT=gs://beam-flink-cluster/init-actions/docker.sh
+ FLINK_LOCAL_PORT=8081
+ FLINK_TASKMANAGER_SLOTS=1
+ TASK_MANAGER_MEM=10240
+ YARN_APPLICATION_MASTER=
+ main
+ upload_init_actions
+ echo 'Uploading initialization actions to GCS bucket: gs://beam-flink-cluster'
Uploading initialization actions to GCS bucket: gs://beam-flink-cluster
+ gsutil cp -r init-actions/beam.sh init-actions/docker.sh 
init-actions/flink.sh gs://beam-flink-cluster/init-actions
Copying file://init-actions/beam.sh [Content-Type=text/x-sh]...
/ [0 files][    0.0 B/  2.3 KiB]                                                
/ [1 files][  2.3 KiB/  2.3 KiB]                                                
Copying file://init-actions/docker.sh [Content-Type=text/x-sh]...
/ [1 files][  2.3 KiB/  6.0 KiB]                                                
/ [2 files][  6.0 KiB/  6.0 KiB]                                                
Copying file://init-actions/flink.sh [Content-Type=text/x-sh]...
/ [2 files][  6.0 KiB/ 13.1 KiB]                                                
/ [3 files][ 13.1 KiB/ 13.1 KiB]                                                
-
Operation completed over 3 objects/13.1 KiB.                                    
 
+ create_cluster
+ local 
metadata=flink-snapshot-url=https://archive.apache.org/dist/flink/flink-1.7.0/flink-1.7.0-bin-hadoop28-scala_2.11.tgz,
+ metadata+=flink-start-yarn-session=true,
+ metadata+=flink-taskmanager-slots=1
+ [[ -n gcr.io/apache-beam-testing/beam_portability/python:latest ]]
+ 
metadata+=,beam-sdk-harness-images-to-pull=gcr.io/apache-beam-testing/beam_portability/python:latest
+ [[ -n gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest ]]
+ 
metadata+=,beam-job-server-image=gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest
+ local image_version=1.2
+ echo 'Starting dataproc cluster. Dataproc version: 1.2'
Starting dataproc cluster. Dataproc version: 1.2
+ local num_dataproc_workers=6
+ gcloud dataproc clusters create beam-loadtests-python-gbk-flink-batch-8 
--num-workers=6 --initialization-actions 
gs://beam-flink-cluster/init-actions/docker.sh,gs://beam-flink-cluster/init-actions/beam.sh,gs://beam-flink-cluster/init-actions/flink.sh
 --metadata 
flink-snapshot-url=https://archive.apache.org/dist/flink/flink-1.7.0/flink-1.7.0-bin-hadoop28-scala_2.11.tgz,flink-start-yarn-session=true,flink-taskmanager-slots=1,beam-sdk-harness-images-to-pull=gcr.io/apache-beam-testing/beam_portability/python:latest,beam-job-server-image=gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest,
 --image-version=1.2 --zone=us-central1-a --quiet
Waiting on operation 
[projects/apache-beam-testing/regions/global/operations/de21d7af-fbd5-3f94-b5c3-791ce036795f].
Waiting for cluster creation operation...
WARNING: For PD-Standard without local SSDs, we strongly recommend provisioning 
1TB or larger to ensure consistently high I/O performance. See 
https://cloud.google.com/compute/docs/disks/performance for information on disk 
I/O performance.
....................................................................................................................................................................................................................................................................................done.
Created 
[https://dataproc.googleapis.com/v1/projects/apache-beam-testing/regions/global/clusters/beam-loadtests-python-gbk-flink-batch-8]
 Cluster placed in zone [us-central1-a].
+ get_leader
+ local i=0
+ local application_ids
+ local application_masters
+ echo 'Yarn Applications'
Yarn Applications
++ gcloud compute ssh --zone=us-central1-a --quiet 
yarn@beam-loadtests-python-gbk-flink-batch-8-m '--command=yarn application 
-list'
++ grep beam-loadtests-python-gbk-flink-batch-8
Warning: Permanently added 'compute.5518065372976524972' (ECDSA) to the list of 
known hosts.
19/06/20 13:59:54 INFO client.RMProxy: Connecting to ResourceManager at 
beam-loadtests-python-gbk-flink-batch-8-m/10.128.0.203:8032
+ read line
+ echo

++ echo
++ sed 's/ .*//'
+ application_ids[$i]=
++ echo
++ sed 
's/.*beam-loadtests-python-gbk-flink-batch-8/beam-loadtests-python-gbk-flink-batch-8/'
++ sed 's/ .*//'
+ application_masters[$i]=
+ i=1
+ read line
+ '[' 1 '!=' 1 ']'
+ YARN_APPLICATION_MASTER=
+ echo 'Using Yarn Application master: '
Using Yarn Application master: 
+ [[ -n gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest ]]
+ start_job_server
+ gcloud compute ssh --zone=us-central1-a --quiet 
yarn@beam-loadtests-python-gbk-flink-batch-8-m '--command=sudo --user yarn 
docker run --detach --publish 8099:8099 --publish 8098:8098 --publish 8097:8097 
--volume ~/.config/gcloud:/root/.config/gcloud 
gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest 
--flink-master-url= 
--artifacts-dir=gs://beam-flink-cluster/beam-loadtests-python-gbk-flink-batch-8'
6cc34401fb3371846ab53a8cef193b3ea1d6a00cd267a62423c02e08e3a09769
+ start_tunnel
++ gcloud compute ssh --quiet --zone=us-central1-a 
yarn@beam-loadtests-python-gbk-flink-batch-8-m '--command=curl -s 
"http:///jobmanager/config";'
+ local job_server_config=
+ local key=jobmanager.rpc.port
++ echo
++ cut -d : -f1
+ local yarn_application_master_host=
++ echo
++ python -c 'import sys, json; print [ e['\''value'\''] for e in 
json.load(sys.stdin) if e['\''key'\''] == u'\''jobmanager.rpc.port'\''][0]'
Traceback (most recent call last):
  File "<string>", line 1, in <module>
  File "/usr/lib/python2.7/json/__init__.py", line 291, in load
    **kw)
  File "/usr/lib/python2.7/json/__init__.py", line 339, in loads
    return _default_decoder.decode(s)
  File "/usr/lib/python2.7/json/decoder.py", line 364, in decode
    obj, end = self.raw_decode(s, idx=_w(s, 0).end())
  File "/usr/lib/python2.7/json/decoder.py", line 382, in raw_decode
    raise ValueError("No JSON object could be decoded")
ValueError: No JSON object could be decoded
+ local jobmanager_rpc_port=
++ [[ true == \t\r\u\e ]]
++ echo ' -Nf >& /dev/null'
+ local 'detached_mode_params= -Nf >& /dev/null'
++ [[ -n gcr.io/apache-beam-testing/beam_portability/flink-job-server:latest ]]
++ echo '-L 8099:localhost:8099 -L 8098:localhost:8098 -L 8097:localhost:8097'
+ local 'job_server_ports_forwarding=-L 8099:localhost:8099 -L 
8098:localhost:8098 -L 8097:localhost:8097'
+ local 'tunnel_command=gcloud compute ssh --zone=us-central1-a --quiet 
yarn@beam-loadtests-python-gbk-flink-batch-8-m -- -L 8081: -L :: -L 
8099:localhost:8099 -L 8098:localhost:8098 -L 8097:localhost:8097 -D 1080  -Nf 
>& /dev/null'
+ eval gcloud compute ssh --zone=us-central1-a --quiet 
yarn@beam-loadtests-python-gbk-flink-batch-8-m -- -L 8081: -L :: -L 
8099:localhost:8099 -L 8098:localhost:8098 -L 8097:localhost:8097 -D 1080 -Nf 
'>&' /dev/null
++ gcloud compute ssh --zone=us-central1-a --quiet 
yarn@beam-loadtests-python-gbk-flink-batch-8-m -- -L 8081: -L :: -L 
8099:localhost:8099 -L 8098:localhost:8098 -L 8097:localhost:8097 -D 1080 -Nf
Build step 'Execute shell' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to