See 
<https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark_Gradle/346/display/redirect?page=changes>

Changes:

[kedin] [SQL] Change DDL parser to create Beam Schema types instead of Calcite

[kedin] [SQL] Support arrays of primitive types in DDL

[kedin] [SQL] Parse Row fields

[kedin] [SQL] Remove Column. Replace with Schema

[kedin] [SQL] Validate Table schemas in BeamSqlCliTest

[kedin] [SQL] Add support for angular brackets array definition

[kedin] [SQL] Add support for Maps in DDL

[kedin] [SQL] Add support for angular brackets Row syntax

[kedin] [SQL] Fix rebase conflicts

[kedin] [SQL] Add QuickCheck and tests for DDL schema verification

[kedin] [SQL] Remove support for postfix array declaration in DDL

------------------------------------------
[...truncated 721.61 KB...]
        at 
org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:747)
        at 
scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
        at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
        at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
        at 
scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:236)
        at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
        at scala.collection.mutable.HashMap.foreach(HashMap.scala:130)
        at 
scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
        at 
org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:747)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:312)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
    2018-05-09 19:01:47,657 [Executor task launch worker for task 16] ERROR 
org.apache.spark.executor.Executor  - Exception in task 0.0 in stage 34.0 (TID 
16)
    java.io.IOException: No such file or directory
        at java.io.UnixFileSystem.createFileExclusively(Native Method)
        at java.io.File.createTempFile(File.java:2024)
        at org.apache.spark.util.Utils$.downloadFile(Utils.scala:527)
        at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:663)
        at org.apache.spark.util.Utils$.fetchFile(Utils.scala:485)
        at 
org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:755)
        at 
org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:747)
        at 
scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
        at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
        at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
        at 
scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:236)
        at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
        at scala.collection.mutable.HashMap.foreach(HashMap.scala:130)
        at 
scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
        at 
org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:747)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:312)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
    2018-05-09 19:01:47,659 [task-result-getter-0] ERROR 
org.apache.spark.scheduler.TaskSetManager  - Task 0 in stage 34.0 failed 1 
times; aborting job
    2018-05-09 19:01:47,661 [JobScheduler] ERROR 
org.apache.spark.streaming.scheduler.JobScheduler  - Error running job 
streaming job 1525892507500 ms.0
    org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 
in stage 34.0 failed 1 times, most recent failure: Lost task 0.0 in stage 34.0 
(TID 16, localhost, executor driver): java.io.IOException: No such file or 
directory
        at java.io.UnixFileSystem.createFileExclusively(Native Method)
        at java.io.File.createTempFile(File.java:2024)
        at org.apache.spark.util.Utils$.downloadFile(Utils.scala:527)
        at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:663)
        at org.apache.spark.util.Utils$.fetchFile(Utils.scala:485)
        at 
org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:755)
        at 
org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:747)
        at 
scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
        at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
        at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
        at 
scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:236)
        at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
        at scala.collection.mutable.HashMap.foreach(HashMap.scala:130)
        at 
scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
        at 
org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:747)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:312)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)

    Driver stacktrace:
        at 
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1599)
        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1587)
        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1586)
        at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
        at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
        at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1586)
        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:831)
        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:831)
        at scala.Option.foreach(Option.scala:257)
        at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:831)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1820)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1769)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1758)
        at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
        at 
org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:642)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2027)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2048)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2067)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2092)
        at org.apache.spark.rdd.RDD$$anonfun$foreach$1.apply(RDD.scala:921)
        at org.apache.spark.rdd.RDD$$anonfun$foreach$1.apply(RDD.scala:919)
        at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
        at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
        at org.apache.spark.rdd.RDD.withScope(RDD.scala:363)
        at org.apache.spark.rdd.RDD.foreach(RDD.scala:919)
        at 
org.apache.spark.api.java.JavaRDDLike$class.foreach(JavaRDDLike.scala:351)
        at 
org.apache.spark.api.java.AbstractJavaRDDLike.foreach(JavaRDDLike.scala:45)
        at 
org.apache.beam.runners.spark.translation.streaming.UnboundedDataset.lambda$action$e3b46054$1(UnboundedDataset.java:81)
        at 
org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$1.apply(JavaDStreamLike.scala:272)
        at 
org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$1.apply(JavaDStreamLike.scala:272)
        at 
org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:628)
        at 
org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:628)
        at 
org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ForEachDStream.scala:51)
        at 
org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:51)
        at 
org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:51)
        at 
org.apache.spark.streaming.dstream.DStream.createRDDWithLocalProperties(DStream.scala:416)
        at 
org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mcV$sp(ForEachDStream.scala:50)
        at 
org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:50)
        at 
org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:50)
        at scala.util.Try$.apply(Try.scala:192)
        at org.apache.spark.streaming.scheduler.Job.run(Job.scala:39)
        at 
org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply$mcV$sp(JobScheduler.scala:257)
        at 
org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:257)
        at 
org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:257)
        at scala.util.DynamicVariable.withValue(DynamicVariable.scala:58)
        at 
org.apache.spark.streaming.scheduler.JobScheduler$JobHandler.run(JobScheduler.scala:256)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
    Caused by: java.io.IOException: No such file or directory
        at java.io.UnixFileSystem.createFileExclusively(Native Method)
        at java.io.File.createTempFile(File.java:2024)
        at org.apache.spark.util.Utils$.downloadFile(Utils.scala:527)
        at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:663)
        at org.apache.spark.util.Utils$.fetchFile(Utils.scala:485)
        at 
org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:755)
        at 
org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:747)
        at 
scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
        at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
        at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
        at 
scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:236)
        at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
        at scala.collection.mutable.HashMap.foreach(HashMap.scala:130)
        at 
scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
        at 
org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:747)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:312)
        ... 3 more
    2018-05-09 19:01:47,674 [Executor task launch worker for task 17] ERROR 
org.apache.spark.executor.Executor  - Exception in task 1.0 in stage 34.0 (TID 
17)
    java.io.IOException: No such file or directory
        at java.io.UnixFileSystem.createFileExclusively(Native Method)
        at java.io.File.createTempFile(File.java:2024)
        at org.apache.spark.util.Utils$.downloadFile(Utils.scala:527)
        at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:663)
        at org.apache.spark.util.Utils$.fetchFile(Utils.scala:485)
        at 
org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:755)
        at 
org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:747)
        at 
scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
        at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
        at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
        at 
scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:236)
        at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
        at scala.collection.mutable.HashMap.foreach(HashMap.scala:130)
        at 
scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
        at 
org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:747)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:312)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
    2018-05-09 19:01:47,689 [Executor task launch worker for task 19] ERROR 
org.apache.spark.executor.Executor  - Exception in task 3.0 in stage 34.0 (TID 
19)
    java.io.IOException: No such file or directory
        at java.io.UnixFileSystem.createFileExclusively(Native Method)
        at java.io.File.createTempFile(File.java:2024)
        at org.apache.spark.util.Utils$.downloadFile(Utils.scala:527)
        at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:663)
        at org.apache.spark.util.Utils$.fetchFile(Utils.scala:485)
        at 
org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:755)
        at 
org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:747)
        at 
scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
        at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
        at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
        at 
scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:236)
        at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
        at scala.collection.mutable.HashMap.foreach(HashMap.scala:130)
        at 
scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
        at 
org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:747)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:312)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
    2018-05-09 19:01:47,704 [Executor task launch worker for task 18] ERROR 
org.apache.spark.executor.Executor  - Exception in task 2.0 in stage 34.0 (TID 
18)
    java.io.IOException: No such file or directory
        at java.io.UnixFileSystem.createFileExclusively(Native Method)
        at java.io.File.createTempFile(File.java:2024)
        at org.apache.spark.util.Utils$.downloadFile(Utils.scala:527)
        at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:663)
        at org.apache.spark.util.Utils$.fetchFile(Utils.scala:485)
        at 
org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:755)
        at 
org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:747)
        at 
scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
        at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
        at 
scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
        at 
scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:236)
        at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
        at scala.collection.mutable.HashMap.foreach(HashMap.scala:130)
        at 
scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
        at 
org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:747)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:312)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
    2018-05-09 19:01:51,647 [rpc-server-24-2] ERROR 
org.apache.spark.network.server.TransportRequestHandler  - Error sending result 
StreamResponse{streamId=/jars/scala-compiler-2.11.0.jar, byteCount=13655547, 
body=FileSegmentManagedBuffer{file=/home/jenkins/.m2/repository/org/scala-lang/scala-compiler/2.11.0/scala-compiler-2.11.0.jar,
 offset=0, length=13655547}} to /127.0.0.1:39314; closing connection
    java.nio.channels.ClosedChannelException
        at io.netty.channel.AbstractChannel$AbstractUnsafe.close(...)(Unknown 
Source)

org.apache.beam.runners.spark.translation.streaming.CreateStreamTest > 
testElementsAtAlmostPositiveInfinity FAILED
    org.apache.beam.sdk.Pipeline$PipelineExecutionException: 
java.io.IOException: No such file or directory
        at 
org.apache.beam.runners.spark.SparkPipelineResult.beamExceptionFrom(SparkPipelineResult.java:68)
        at 
org.apache.beam.runners.spark.SparkPipelineResult.access$000(SparkPipelineResult.java:41)
        at 
org.apache.beam.runners.spark.SparkPipelineResult$StreamingMode.stop(SparkPipelineResult.java:163)
        at 
org.apache.beam.runners.spark.TestSparkRunner.run(TestSparkRunner.java:125)
        at 
org.apache.beam.runners.spark.TestSparkRunner.run(TestSparkRunner.java:83)
        at org.apache.beam.sdk.Pipeline.run(Pipeline.java:311)
        at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:346)
        at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:328)
        at 
org.apache.beam.runners.spark.translation.streaming.CreateStreamTest.testElementsAtAlmostPositiveInfinity(CreateStreamTest.java:265)

        Caused by:
        java.io.IOException: No such file or directory

org.apache.beam.runners.spark.translation.streaming.CreateStreamTest > 
testFirstElementLate STANDARD_ERROR
    2018-05-09 19:02:01,682 [Test worker] INFO  
org.apache.beam.runners.spark.TestSparkRunner  - About to run test pipeline 
createstreamtest0testfirstelementlate-jenkins-0509190201-6dc4de8

org.apache.beam.runners.spark.translation.streaming.CreateStreamTest > 
testDiscardingMode STANDARD_ERROR
    2018-05-09 19:02:06,932 [Test worker] INFO  
org.apache.beam.runners.spark.TestSparkRunner  - About to run test pipeline 
createstreamtest0testdiscardingmode-jenkins-0509190206-a0afd5a6

Gradle Test Executor 288 finished executing tests.

> Task :beam-runners-spark:validatesRunnerStreaming FAILED

13 tests completed, 1 failed
Finished generating test XML results (0.005 secs) into: 
<https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark_Gradle/ws/src/runners/spark/build/test-results/validatesRunnerStreaming>
Generating HTML test report...
Finished generating test html results (0.01 secs) into: 
<https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark_Gradle/ws/src/runners/spark/build/reports/tests/validatesRunnerStreaming>
:beam-runners-spark:validatesRunnerStreaming (Thread[Task worker for ':' Thread 
8,5,main]) completed. Took 1 mins 18.705 secs.

FAILURE: Build failed with an exception.

* What went wrong:
Execution failed for task ':beam-runners-spark:validatesRunnerStreaming'.
> There were failing tests. See the report at: 
> file://<https://builds.apache.org/job/beam_PostCommit_Java_ValidatesRunner_Spark_Gradle/ws/src/runners/spark/build/reports/tests/validatesRunnerStreaming/index.html>

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to 
get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with 
Gradle 5.0.
See 
https://docs.gradle.org/4.7/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 5m 4s
39 actionable tasks: 36 executed, 3 from cache

Publishing build scan...
https://gradle.com/s/7xou44udvg4pu

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
Recording test results
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user 
[email protected]
Not sending mail to unregistered user [email protected]

Reply via email to