See 
<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/222/display/redirect>

------------------------------------------
[...truncated 18.62 MB...]

org.apache.beam.sdk.io.gcp.datastore.V1ReadIT > 
testE2EV1ReadWithGQLQueryWithNoLimit STANDARD_ERROR
    Apr 28, 2018 12:45:21 PM 
org.apache.beam.runners.dataflow.DataflowPipelineJob waitUntilFinish
    INFO: Job 2018-04-28_05_42_10-3656998925029165899 finished with status DONE.
    Apr 28, 2018 12:45:21 PM 
org.apache.beam.runners.dataflow.TestDataflowRunner checkForPAssertSuccess
    INFO: Success result for Dataflow job 
2018-04-28_05_42_10-3656998925029165899. Found 1 success, 0 failures out of 1 
expected assertions.
    Apr 28, 2018 12:45:22 PM 
org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch
    INFO: Writing batch of 500 entities
    Apr 28, 2018 12:45:22 PM 
org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch
    INFO: Successfully wrote 500 entities
    Apr 28, 2018 12:45:24 PM 
org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch
    INFO: Writing batch of 500 entities
    Apr 28, 2018 12:45:24 PM 
org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch
    INFO: Successfully wrote 500 entities
    Apr 28, 2018 12:45:24 PM org.apache.beam.sdk.io.gcp.datastore.V1TestUtil 
deleteAllEntities
    INFO: Successfully deleted 1000 entities

org.apache.beam.sdk.io.gcp.datastore.SplitQueryFnIT > 
testSplitQueryFnWithLargeDataset STANDARD_ERROR
    Apr 28, 2018 12:45:24 PM org.apache.beam.sdk.transforms.DoFnTester of
    WARNING: Your tests use DoFnTester, which may not exercise DoFns correctly. 
Please use TestPipeline instead.
    Apr 28, 2018 12:45:24 PM 
org.apache.beam.sdk.io.gcp.datastore.DatastoreV1$Read getEstimatedSizeBytes
    INFO: Latest stats timestamp for kind sort_1G is 1524728602000000
    Apr 28, 2018 12:45:24 PM 
org.apache.beam.sdk.io.gcp.datastore.DatastoreV1$Read getEstimatedNumSplits
    INFO: Estimated size bytes for the query is: 2130000000
    Apr 28, 2018 12:45:24 PM 
org.apache.beam.sdk.io.gcp.datastore.DatastoreV1$Read$SplitQueryFn 
processElement
    INFO: Splitting the query into 32 splits

org.apache.beam.sdk.io.gcp.datastore.SplitQueryFnIT > 
testSplitQueryFnWithSmallDataset STANDARD_ERROR
    Apr 28, 2018 12:45:25 PM org.apache.beam.sdk.transforms.DoFnTester of
    WARNING: Your tests use DoFnTester, which may not exercise DoFns correctly. 
Please use TestPipeline instead.
    Apr 28, 2018 12:45:25 PM 
org.apache.beam.sdk.io.gcp.datastore.DatastoreV1$Read getEstimatedSizeBytes
    INFO: Latest stats timestamp for kind shakespeare is 1524728602000000
    Apr 28, 2018 12:45:25 PM 
org.apache.beam.sdk.io.gcp.datastore.DatastoreV1$Read getEstimatedNumSplits
    INFO: Estimated size bytes for the query is: 26383451
    Apr 28, 2018 12:45:25 PM 
org.apache.beam.sdk.io.gcp.datastore.DatastoreV1$Read$SplitQueryFn 
processElement
    INFO: Splitting the query into 12 splits

org.apache.beam.sdk.io.gcp.datastore.V1WriteIT > testE2EV1Write STANDARD_ERROR
    Apr 28, 2018 12:45:25 PM org.apache.beam.runners.dataflow.util.PackageUtil 
stageClasspathElements
    INFO: Staging files complete: 0 files cached, 128 files newly uploaded
    Apr 28, 2018 12:45:25 PM 
org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding GenerateSequence/Read(BoundedCountingSource) as step s1
    Apr 28, 2018 12:45:25 PM 
org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding ParDo(CreateEntity) as step s2
    Apr 28, 2018 12:45:25 PM 
org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding DatastoreV1.Write/Convert to Mutation/Map as step s3
    Apr 28, 2018 12:45:25 PM 
org.apache.beam.runners.dataflow.DataflowPipelineTranslator$Translator addStep
    INFO: Adding DatastoreV1.Write/Write Mutation to Datastore as step s4
    Apr 28, 2018 12:45:25 PM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: Staging pipeline description to 
gs://temp-storage-for-end-to-end-tests/testpipeline-jenkins-0428124518-c2b255f4/output/results/staging/
    Apr 28, 2018 12:45:25 PM org.apache.beam.runners.dataflow.util.PackageUtil 
tryStagePackage
    INFO: Uploading <8774 bytes, hash faqaIfuEyloIF_lTeB6mtA> to 
gs://temp-storage-for-end-to-end-tests/testpipeline-jenkins-0428124518-c2b255f4/output/results/staging/pipeline-faqaIfuEyloIF_lTeB6mtA.pb

org.apache.beam.sdk.io.gcp.datastore.V1WriteIT > testE2EV1Write STANDARD_OUT
    Dataflow SDK version: 2.5.0-SNAPSHOT

Gradle Test Executor 121 finished executing tests.

> Task 
> :beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest

org.apache.beam.sdk.io.gcp.spanner.SpannerWriteIT > testWrite FAILED
    java.lang.NoClassDefFoundError: 
com/google/api/gax/retrying/ExceptionRetryAlgorithm
        at java.lang.ClassLoader.defineClass1(Native Method)
        at java.lang.ClassLoader.defineClass(ClassLoader.java:763)
        at 
java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
        at java.net.URLClassLoader.defineClass(URLClassLoader.java:467)
        at java.net.URLClassLoader.access$100(URLClassLoader.java:73)
        at java.net.URLClassLoader$1.run(URLClassLoader.java:368)
        at java.net.URLClassLoader$1.run(URLClassLoader.java:362)
        at java.security.AccessController.doPrivileged(Native Method)
        at java.net.URLClassLoader.findClass(URLClassLoader.java:361)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
        at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:338)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
        at com.google.cloud.BaseService.<clinit>(BaseService.java:48)
        at 
com.google.cloud.spanner.SpannerOptions$DefaultSpannerFactory.create(SpannerOptions.java:60)
        at 
com.google.cloud.spanner.SpannerOptions$DefaultSpannerFactory.create(SpannerOptions.java:55)
        at com.google.cloud.ServiceOptions.getService(ServiceOptions.java:426)
        at 
org.apache.beam.sdk.io.gcp.spanner.SpannerWriteIT.setUp(SpannerWriteIT.java:91)

        Caused by:
        java.lang.ClassNotFoundException: 
com.google.api.gax.retrying.ExceptionRetryAlgorithm
            at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
            at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
            at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:338)
            at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
            ... 17 more

    java.lang.NullPointerException
        at 
org.apache.beam.sdk.io.gcp.spanner.SpannerWriteIT.tearDown(SpannerWriteIT.java:148)

org.apache.beam.sdk.io.gcp.datastore.V1WriteIT > testE2EV1Write STANDARD_ERROR
    Apr 28, 2018 12:45:26 PM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: To access the Dataflow monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-04-28_05_45_26-14475005888636693470?project=apache-beam-testing

org.apache.beam.sdk.io.gcp.datastore.V1WriteIT > testE2EV1Write STANDARD_OUT
    Submitted job: 2018-04-28_05_45_26-14475005888636693470

org.apache.beam.sdk.io.gcp.datastore.V1WriteIT > testE2EV1Write STANDARD_ERROR
    Apr 28, 2018 12:45:26 PM org.apache.beam.runners.dataflow.DataflowRunner run
    INFO: To cancel the job using the 'gcloud' tool, run:
    > gcloud dataflow jobs --project=apache-beam-testing cancel 
--region=us-central1 2018-04-28_05_45_26-14475005888636693470
    Apr 28, 2018 12:45:26 PM 
org.apache.beam.runners.dataflow.TestDataflowRunner run
    INFO: Running Dataflow job 2018-04-28_05_45_26-14475005888636693470 with 0 
expected assertions.
    Apr 28, 2018 12:45:38 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:45:26.213Z: Autoscaling is enabled for job 
2018-04-28_05_45_26-14475005888636693470. The number of workers will be between 
1 and 1000.
    Apr 28, 2018 12:45:38 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:45:26.237Z: Autoscaling was automatically enabled for 
job 2018-04-28_05_45_26-14475005888636693470.
    Apr 28, 2018 12:45:38 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:45:28.605Z: Checking required Cloud APIs are enabled.
    Apr 28, 2018 12:45:38 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:45:28.778Z: Checking permissions granted to controller 
Service Account.
    Apr 28, 2018 12:45:38 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:45:33.670Z: Worker configuration: n1-standard-1 in 
us-central1-f.
    Apr 28, 2018 12:45:38 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:45:34.011Z: Expanding CoGroupByKey operations into 
optimizable parts.
    Apr 28, 2018 12:45:38 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:45:34.045Z: Expanding GroupByKey operations into 
optimizable parts.
    Apr 28, 2018 12:45:38 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:45:34.070Z: Lifting ValueCombiningMappingFns into 
MergeBucketsMappingFns
    Apr 28, 2018 12:45:38 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:45:34.119Z: Fusing adjacent ParDo, Read, Write, and 
Flatten operations
    Apr 28, 2018 12:45:38 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:45:34.145Z: Fusing consumer ParDo(CreateEntity) into 
GenerateSequence/Read(BoundedCountingSource)
    Apr 28, 2018 12:45:38 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:45:34.174Z: Fusing consumer DatastoreV1.Write/Write 
Mutation to Datastore into DatastoreV1.Write/Convert to Mutation/Map
    Apr 28, 2018 12:45:38 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:45:34.198Z: Fusing consumer DatastoreV1.Write/Convert 
to Mutation/Map into ParDo(CreateEntity)
    Apr 28, 2018 12:45:38 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:45:34.473Z: Executing operation 
GenerateSequence/Read(BoundedCountingSource)+ParDo(CreateEntity)+DatastoreV1.Write/Convert
 to Mutation/Map+DatastoreV1.Write/Write Mutation to Datastore
    Apr 28, 2018 12:45:38 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:45:34.536Z: Starting 1 workers in us-central1-f...
    Apr 28, 2018 12:45:45 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:45:43.918Z: Autoscaling: Raised the number of workers 
to 0 based on the rate of progress in the currently running step(s).
    Apr 28, 2018 12:46:10 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:46:10.252Z: Autoscaling: Raised the number of workers 
to 1 based on the rate of progress in the currently running step(s).
    Apr 28, 2018 12:46:33 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:46:32.148Z: Workers have started successfully.
    Apr 28, 2018 12:47:00 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:47:00.106Z: Cleaning up.
    Apr 28, 2018 12:47:00 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:47:00.175Z: Stopping worker pool...
    Apr 28, 2018 12:48:26 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:48:25.822Z: Autoscaling: Resized worker pool from 1 to 
0.
    Apr 28, 2018 12:48:26 PM 
org.apache.beam.runners.dataflow.util.MonitoringUtil$LoggingHandler process
    INFO: 2018-04-28T12:48:25.849Z: Autoscaling: Would further reduce the 
number of workers but reached the minimum number allowed for the job.
    Apr 28, 2018 12:48:33 PM 
org.apache.beam.runners.dataflow.DataflowPipelineJob waitUntilFinish
    INFO: Job 2018-04-28_05_45_26-14475005888636693470 finished with status 
DONE.
    Apr 28, 2018 12:48:33 PM 
org.apache.beam.runners.dataflow.TestDataflowRunner checkForPAssertSuccess
    INFO: Success result for Dataflow job 
2018-04-28_05_45_26-14475005888636693470. Found 0 success, 0 failures out of 0 
expected assertions.
    Apr 28, 2018 12:48:35 PM 
org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch
    INFO: Writing batch of 500 entities
    Apr 28, 2018 12:48:35 PM 
org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch
    INFO: Successfully wrote 500 entities
    Apr 28, 2018 12:48:35 PM 
org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch
    INFO: Writing batch of 500 entities
    Apr 28, 2018 12:48:36 PM 
org.apache.beam.sdk.io.gcp.datastore.V1TestUtil$V1TestWriter flushBatch
    INFO: Successfully wrote 500 entities
    Apr 28, 2018 12:48:36 PM org.apache.beam.sdk.io.gcp.datastore.V1TestUtil 
deleteAllEntities
    INFO: Successfully deleted 1000 entities

org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT > testQuery FAILED
    java.lang.NoClassDefFoundError: 
com/google/api/gax/retrying/ExceptionRetryAlgorithm
        at java.lang.ClassLoader.defineClass1(Native Method)
        at java.lang.ClassLoader.defineClass(ClassLoader.java:763)
        at 
java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142)
        at java.net.URLClassLoader.defineClass(URLClassLoader.java:467)
        at java.net.URLClassLoader.access$100(URLClassLoader.java:73)
        at java.net.URLClassLoader$1.run(URLClassLoader.java:368)
        at java.net.URLClassLoader$1.run(URLClassLoader.java:362)
        at java.security.AccessController.doPrivileged(Native Method)
        at java.net.URLClassLoader.findClass(URLClassLoader.java:361)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
        at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:338)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
        at com.google.cloud.BaseService.<clinit>(BaseService.java:48)
        at 
com.google.cloud.spanner.SpannerOptions$DefaultSpannerFactory.create(SpannerOptions.java:60)
        at 
com.google.cloud.spanner.SpannerOptions$DefaultSpannerFactory.create(SpannerOptions.java:55)
        at com.google.cloud.ServiceOptions.getService(ServiceOptions.java:426)
        at 
org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT.setUp(SpannerReadIT.java:90)

        Caused by:
        java.lang.ClassNotFoundException: 
com.google.api.gax.retrying.ExceptionRetryAlgorithm
            at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
            at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
            at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:338)
            at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
            ... 17 more

    java.lang.NullPointerException
        at 
org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT.tearDown(SpannerReadIT.java:198)

Gradle Test Executor 130 finished executing tests.

> Task 
> :beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest 
> FAILED

org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT > testRead FAILED
    java.lang.NoClassDefFoundError: Could not initialize class 
com.google.cloud.spanner.SpannerImpl
        at 
com.google.cloud.spanner.SpannerOptions$DefaultSpannerFactory.create(SpannerOptions.java:60)
        at 
com.google.cloud.spanner.SpannerOptions$DefaultSpannerFactory.create(SpannerOptions.java:55)
        at com.google.cloud.ServiceOptions.getService(ServiceOptions.java:426)
        at 
org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT.setUp(SpannerReadIT.java:90)

    java.lang.NullPointerException
        at 
org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT.tearDown(SpannerReadIT.java:198)

12 tests completed, 3 failed
Finished generating test XML results (0.007 secs) into: 
<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/test-results/googleCloudPlatformIntegrationTest>
Generating HTML test report...
Finished generating test html results (0.008 secs) into: 
<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/reports/tests/googleCloudPlatformIntegrationTest>
:beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest 
(Thread[main,5,main]) completed. Took 19 mins 5.08 secs.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task 
':beam-runners-google-cloud-dataflow-java:googleCloudPlatformIntegrationTest'.
> There were failing tests. See the report at: 
> file://<https://builds.apache.org/job/beam_PostCommit_Java_GradleBuild/ws/src/runners/google-cloud-dataflow-java/build/reports/tests/googleCloudPlatformIntegrationTest/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to 
get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with 
Gradle 5.0.
See 
https://docs.gradle.org/4.7/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 47m 8s
546 actionable tasks: 541 executed, 5 from cache

Publishing build scan...
https://gradle.com/s/ggn2jjuxipceq

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
Recording test results
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user 
[email protected]
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]

Reply via email to