See
<https://builds.apache.org/job/beam_PostCommit_Go/5648/display/redirect?page=changes>
Changes:
[pawel.pasterz] [BEAM-8978] Publish table size of data written during
HadoopFormatIOIT
------------------------------------------
[...truncated 298.80 KB...]
"@type": "kind:length_prefix",
"component_encodings": [
{
"@type":
"Cgd2YXJpbnR6EgIIAhqFAQpxZ2l0aHViLmNvbS9hcGFjaGUvYmVhbS9zZGtzL2dvL3Rlc3QvdmVuZG9yL2dpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby9wa2cvYmVhbS9jb3JlL3J1bnRpbWUvY29kZXJ4LmVuY1ZhckludFoSEAgWIgQIGUAPKgYIFBICCAgikQEKcWdpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby90ZXN0L3ZlbmRvci9naXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vcGtnL2JlYW0vY29yZS9ydW50aW1lL2NvZGVyeC5kZWNWYXJJbnRaEhwIFiIECBlAAyIGCBQSAggIKgQIGUAPKgQIGUAB"
}
]
},
{
"@type": "kind:length_prefix",
"component_encodings": [
{
"@type":
"Cgd2YXJpbnR6EgIIAhqFAQpxZ2l0aHViLmNvbS9hcGFjaGUvYmVhbS9zZGtzL2dvL3Rlc3QvdmVuZG9yL2dpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby9wa2cvYmVhbS9jb3JlL3J1bnRpbWUvY29kZXJ4LmVuY1ZhckludFoSEAgWIgQIGUAPKgYIFBICCAgikQEKcWdpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby90ZXN0L3ZlbmRvci9naXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vcGtnL2JlYW0vY29yZS9ydW50aW1lL2NvZGVyeC5kZWNWYXJJbnRaEhwIFiIECBlAAyIGCBQSAggIKgQIGUAPKgQIGUAB"
}
]
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
}
}
],
"parallel_input": {
"@type": "OutputReference",
"step_name": "e4",
"output_name": "i0"
},
"serialized_fn": "e5"
}
},
{
"kind": "GroupByKey",
"name": "e6",
"properties": {
"user_name": "passert.Sum(out)/CoGBK",
"output_info": [
{
"user_name": "i0",
"output_name": "i0",
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type": "kind:pair",
"component_encodings": [
{
"@type": "kind:length_prefix",
"component_encodings": [
{
"@type":
"Cgd2YXJpbnR6EgIIAhqFAQpxZ2l0aHViLmNvbS9hcGFjaGUvYmVhbS9zZGtzL2dvL3Rlc3QvdmVuZG9yL2dpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby9wa2cvYmVhbS9jb3JlL3J1bnRpbWUvY29kZXJ4LmVuY1ZhckludFoSEAgWIgQIGUAPKgYIFBICCAgikQEKcWdpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby90ZXN0L3ZlbmRvci9naXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vcGtnL2JlYW0vY29yZS9ydW50aW1lL2NvZGVyeC5kZWNWYXJJbnRaEhwIFiIECBlAAyIGCBQSAggIKgQIGUAPKgQIGUAB"
}
]
},
{
"@type": "kind:stream",
"component_encodings": [
{
"@type": "kind:length_prefix",
"component_encodings": [
{
"@type":
"Cgd2YXJpbnR6EgIIAhqFAQpxZ2l0aHViLmNvbS9hcGFjaGUvYmVhbS9zZGtzL2dvL3Rlc3QvdmVuZG9yL2dpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby9wa2cvYmVhbS9jb3JlL3J1bnRpbWUvY29kZXJ4LmVuY1ZhckludFoSEAgWIgQIGUAPKgYIFBICCAgikQEKcWdpdGh1Yi5jb20vYXBhY2hlL2JlYW0vc2Rrcy9nby90ZXN0L3ZlbmRvci9naXRodWIuY29tL2FwYWNoZS9iZWFtL3Nka3MvZ28vcGtnL2JlYW0vY29yZS9ydW50aW1lL2NvZGVyeC5kZWNWYXJJbnRaEhwIFiIECBlAAyIGCBQSAggIKgQIGUAPKgQIGUAB"
}
]
}
],
"is_stream_like": true
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
}
}
],
"parallel_input": {
"@type": "OutputReference",
"step_name": "e5",
"output_name": "i0"
},
"serialized_fn":
"%0A%27%22%25%0A%02c1%12%1F%0A%1D%0A%1Bbeam:coder:global_window:v1j7%0A%23%0A%21beam:windowfn:global_windows:v0.1%10%01%1A%02c1%22%02:%00%28%010%018%02H%01"
}
},
{
"kind": "ParallelDo",
"name": "e7",
"properties": {
"user_name": "passert.Sum(out)/passert.sumFn",
"output_info": [
{
"user_name": "bogus",
"output_name": "bogus",
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type": "kind:bytes"
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
}
}
],
"parallel_input": {
"@type": "OutputReference",
"step_name": "e6",
"output_name": "i0"
},
"serialized_fn": "e7"
}
}
],
"type": "JOB_TYPE_BATCH"
}
2019/12/20 09:31:07 Staged model pipeline:
gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/go-6-1576834262841220837/model
2019/12/20 09:31:07 Test pardo:multioutput failed: Dataflow supports one
container image only: []
2019/12/20 09:31:08 Submitted job: 2019-12-20_01_31_06-3570159288780251086
2019/12/20 09:31:08 Console:
https://console.cloud.google.com/dataflow/job/2019-12-20_01_31_06-3570159288780251086?project=apache-beam-testing
2019/12/20 09:31:08 Logs:
https://console.cloud.google.com/logs/viewer?project=apache-beam-testing&resource=dataflow_step%2Fjob_id%2F2019-12-20_01_31_06-3570159288780251086
2019/12/20 09:31:08 Submitted job: 2019-12-20_01_31_06-8660428734466252032
2019/12/20 09:31:08 Console:
https://console.cloud.google.com/dataflow/job/2019-12-20_01_31_06-8660428734466252032?project=apache-beam-testing
2019/12/20 09:31:08 Logs:
https://console.cloud.google.com/logs/viewer?project=apache-beam-testing&resource=dataflow_step%2Fjob_id%2F2019-12-20_01_31_06-8660428734466252032
2019/12/20 09:31:08 Job state: JOB_STATE_PENDING ...
2019/12/20 09:31:08 Job state: JOB_STATE_PENDING ...
2019/12/20 09:31:08 Submitted job: 2019-12-20_01_31_07-1795504959596964250
2019/12/20 09:31:08 Console:
https://console.cloud.google.com/dataflow/job/2019-12-20_01_31_07-1795504959596964250?project=apache-beam-testing
2019/12/20 09:31:08 Logs:
https://console.cloud.google.com/logs/viewer?project=apache-beam-testing&resource=dataflow_step%2Fjob_id%2F2019-12-20_01_31_07-1795504959596964250
2019/12/20 09:31:08 Submitted job: 2019-12-20_01_31_07-12678198986439051365
2019/12/20 09:31:08 Console:
https://console.cloud.google.com/dataflow/job/2019-12-20_01_31_07-12678198986439051365?project=apache-beam-testing
2019/12/20 09:31:08 Logs:
https://console.cloud.google.com/logs/viewer?project=apache-beam-testing&resource=dataflow_step%2Fjob_id%2F2019-12-20_01_31_07-12678198986439051365
2019/12/20 09:31:08 Job state: JOB_STATE_PENDING ...
2019/12/20 09:31:08 Submitted job: 2019-12-20_01_31_07-15644918244736508210
2019/12/20 09:31:08 Console:
https://console.cloud.google.com/dataflow/job/2019-12-20_01_31_07-15644918244736508210?project=apache-beam-testing
2019/12/20 09:31:08 Logs:
https://console.cloud.google.com/logs/viewer?project=apache-beam-testing&resource=dataflow_step%2Fjob_id%2F2019-12-20_01_31_07-15644918244736508210
2019/12/20 09:31:09 Job state: JOB_STATE_PENDING ...
2019/12/20 09:31:09 Job state: JOB_STATE_PENDING ...
2019/12/20 09:31:38 Job still running ...
2019/12/20 09:31:38 Job still running ...
2019/12/20 09:31:38 Job still running ...
2019/12/20 09:31:39 Job still running ...
2019/12/20 09:31:39 Job still running ...
2019/12/20 09:32:08 Job still running ...
2019/12/20 09:32:08 Job still running ...
2019/12/20 09:32:08 Job still running ...
2019/12/20 09:32:09 Job still running ...
2019/12/20 09:32:09 Job still running ...
2019/12/20 09:32:38 Job still running ...
2019/12/20 09:32:38 Job still running ...
2019/12/20 09:32:38 Job still running ...
2019/12/20 09:32:39 Job still running ...
2019/12/20 09:32:39 Job still running ...
2019/12/20 09:33:08 Job still running ...
2019/12/20 09:33:08 Job still running ...
2019/12/20 09:33:09 Job still running ...
2019/12/20 09:33:09 Job still running ...
2019/12/20 09:33:09 Job still running ...
2019/12/20 09:33:38 Job still running ...
2019/12/20 09:33:38 Job still running ...
2019/12/20 09:33:39 Job still running ...
2019/12/20 09:33:39 Job still running ...
2019/12/20 09:33:39 Job still running ...
2019/12/20 09:34:08 Job succeeded!
2019/12/20 09:34:08 Test pardo:kvsideinput completed
2019/12/20 09:34:09 Job succeeded!
2019/12/20 09:34:09 Test wordcount:memfs completed
2019/12/20 09:34:09 Job still running ...
2019/12/20 09:34:09 Job succeeded!
2019/12/20 09:34:09 Test wordcount:kinglear completed
2019/12/20 09:34:09 Job still running ...
2019/12/20 09:34:39 Job succeeded!
2019/12/20 09:34:39 Test cogbk:cogbk completed
2019/12/20 09:34:39 Job succeeded!
2019/12/20 09:34:39 Test pardo:sideinput completed
2019/12/20 09:34:39 Result: 2 tests failed
if [[ ! -z "$JOB_PORT" ]]; then
# Shut down the job server
kill %1 || echo "Failed to shut down job server"
fi
# Delete the container locally and remotely
docker rmi $CONTAINER:$TAG || echo "Failed to remove container"
Untagged: us.gcr.io/apache-beam-testing/jenkins/go_sdk:20191220-092935
Untagged:
us.gcr.io/apache-beam-testing/jenkins/go_sdk@sha256:c872c3858f43aa163bf655f28f695bd1ce79ce1042004940ff60a7ffa355ff18
Deleted: sha256:f0b00c13d942614e73f8481f4471fe11b18a3d16d2095bd5deb7811c65e3c437
Deleted: sha256:f11f35317fedb44f40217fa0f42e051b728c2fee5ffb49622dd65694f83be366
Deleted: sha256:c2e3327f60920615d100741030f86a37d3e0d22cb251bcc5bab9be2cdb4a05ce
Deleted: sha256:97624f5555ea5e5e1e92ccca716d3e7a553e0504c3b4924a156cd1147c0de4ae
Deleted: sha256:a95bd6bf9e80744c274a69c608c4cfc9e65441e6b3c5bf3a3d8acb353dccf60b
Deleted: sha256:57c6bd038f8e244a91fbaaac1ddfa561b32debec5370e1d9600b320f9f379de8
gcloud --quiet container images delete $CONTAINER:$TAG || echo "Failed to
delete container"
Digests:
-
us.gcr.io/apache-beam-testing/jenkins/go_sdk@sha256:c872c3858f43aa163bf655f28f695bd1ce79ce1042004940ff60a7ffa355ff18
Associated tags:
- 20191220-092935
Tags:
- us.gcr.io/apache-beam-testing/jenkins/go_sdk:20191220-092935
Deleted [us.gcr.io/apache-beam-testing/jenkins/go_sdk:20191220-092935].
Deleted
[us.gcr.io/apache-beam-testing/jenkins/go_sdk@sha256:c872c3858f43aa163bf655f28f695bd1ce79ce1042004940ff60a7ffa355ff18].
# Clean up tempdir
rm -rf $TMPDIR
if [[ "$TEST_EXIT_CODE" -eq 0 ]]; then
echo ">>> SUCCESS"
else
echo ">>> FAILURE"
fi
exit $TEST_EXIT_CODE
>>> FAILURE
> Task :goIntegrationTests FAILED
FAILURE: Build failed with an exception.
* Where:
Build file
'<https://builds.apache.org/job/beam_PostCommit_Go/ws/src/build.gradle'> line:
195
* What went wrong:
Execution failed for task ':goIntegrationTests'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug
option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with
Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See
https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 8m 11s
72 actionable tasks: 55 executed, 17 from cache
Publishing build scan...
https://gradle.com/s/mdl2thd4q5hau
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]