See <https://builds.apache.org/job/beam_PostCommit_Java_Nexmark_Spark/1036/display/redirect?page=changes>
Changes: [github] Update built-in.md ------------------------------------------ [...truncated 892.23 KB...] at java.lang.Thread.run(Thread.java:748) 18/11/12 11:27:42 ERROR org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter: Error while deleting file /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/28/temp_shuffle_dc491e52-da7c-4351-86d1-f798c0dbae5a 18/11/12 11:27:42 ERROR org.apache.spark.storage.DiskBlockObjectWriter: Uncaught exception while reverting partial writes to file /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/07/temp_shuffle_7dd78bed-2e4a-43f2-a562-c49f16bf7cb3 java.io.FileNotFoundException: /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/07/temp_shuffle_7dd78bed-2e4a-43f2-a562-c49f16bf7cb3 (No such file or directory) at java.io.FileOutputStream.open0(Native Method) at java.io.FileOutputStream.open(FileOutputStream.java:270) at java.io.FileOutputStream.<init>(FileOutputStream.java:213) at org.apache.spark.storage.DiskBlockObjectWriter$$anonfun$revertPartialWritesAndClose$2.apply$mcV$sp(DiskBlockObjectWriter.scala:217) at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1390) at org.apache.spark.storage.DiskBlockObjectWriter.revertPartialWritesAndClose(DiskBlockObjectWriter.scala:214) at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.stop(BypassMergeSortShuffleWriter.java:237) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:102) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) at org.apache.spark.scheduler.Task.run(Task.scala:109) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 18/11/12 11:27:42 ERROR org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter: Error while deleting file /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/07/temp_shuffle_7dd78bed-2e4a-43f2-a562-c49f16bf7cb3 18/11/12 11:27:42 ERROR org.apache.spark.storage.DiskBlockObjectWriter: Uncaught exception while reverting partial writes to file /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/15/temp_shuffle_43301066-7eab-4f2b-ba0a-6bdd1345d946 java.io.FileNotFoundException: /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/15/temp_shuffle_43301066-7eab-4f2b-ba0a-6bdd1345d946 (No such file or directory) at java.io.FileOutputStream.open0(Native Method) at java.io.FileOutputStream.open(FileOutputStream.java:270) at java.io.FileOutputStream.<init>(FileOutputStream.java:213) at org.apache.spark.storage.DiskBlockObjectWriter$$anonfun$revertPartialWritesAndClose$2.apply$mcV$sp(DiskBlockObjectWriter.scala:217) at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1390) at org.apache.spark.storage.DiskBlockObjectWriter.revertPartialWritesAndClose(DiskBlockObjectWriter.scala:214) at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.stop(BypassMergeSortShuffleWriter.java:237) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:102) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) at org.apache.spark.scheduler.Task.run(Task.scala:109) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 18/11/12 11:27:42 ERROR org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter: Error while deleting file /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/15/temp_shuffle_43301066-7eab-4f2b-ba0a-6bdd1345d946 18/11/12 11:27:42 ERROR org.apache.spark.storage.DiskBlockObjectWriter: Uncaught exception while reverting partial writes to file /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/02/temp_shuffle_7c7fa744-d90e-421b-a46e-32c5c59bc220 java.io.FileNotFoundException: /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/02/temp_shuffle_7c7fa744-d90e-421b-a46e-32c5c59bc220 (No such file or directory) at java.io.FileOutputStream.open0(Native Method) at java.io.FileOutputStream.open(FileOutputStream.java:270) at java.io.FileOutputStream.<init>(FileOutputStream.java:213) at org.apache.spark.storage.DiskBlockObjectWriter$$anonfun$revertPartialWritesAndClose$2.apply$mcV$sp(DiskBlockObjectWriter.scala:217) at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1390) at org.apache.spark.storage.DiskBlockObjectWriter.revertPartialWritesAndClose(DiskBlockObjectWriter.scala:214) at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.stop(BypassMergeSortShuffleWriter.java:237) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:102) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) at org.apache.spark.scheduler.Task.run(Task.scala:109) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 18/11/12 11:27:42 ERROR org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter: Error while deleting file /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/02/temp_shuffle_7c7fa744-d90e-421b-a46e-32c5c59bc220 18/11/12 11:27:42 ERROR org.apache.spark.executor.Executor: Exception in task 46.0 in stage 0.0 (TID 46) java.io.FileNotFoundException: /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/3f/temp_shuffle_bc416c9b-bcb3-4c65-899a-f0df16c0fa6f (No such file or directory) at java.io.FileOutputStream.open0(Native Method) at java.io.FileOutputStream.open(FileOutputStream.java:270) at java.io.FileOutputStream.<init>(FileOutputStream.java:213) at org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:103) at org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:116) at org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:237) at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:151) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) at org.apache.spark.scheduler.Task.run(Task.scala:109) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 18/11/12 11:27:42 ERROR org.apache.spark.executor.Executor: Exception in task 47.0 in stage 0.0 (TID 47) java.io.FileNotFoundException: /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/1d/temp_shuffle_6bec34a8-e407-44d0-af66-4204fb9ea9b9 (No such file or directory) at java.io.FileOutputStream.open0(Native Method) at java.io.FileOutputStream.open(FileOutputStream.java:270) at java.io.FileOutputStream.<init>(FileOutputStream.java:213) at org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:103) at org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:116) at org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:237) at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:151) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) at org.apache.spark.scheduler.Task.run(Task.scala:109) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 18/11/12 11:27:42 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 46.0 in stage 0.0 (TID 46, localhost, executor driver): java.io.FileNotFoundException: /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/3f/temp_shuffle_bc416c9b-bcb3-4c65-899a-f0df16c0fa6f (No such file or directory) at java.io.FileOutputStream.open0(Native Method) at java.io.FileOutputStream.open(FileOutputStream.java:270) at java.io.FileOutputStream.<init>(FileOutputStream.java:213) at org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:103) at org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:116) at org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:237) at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:151) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) at org.apache.spark.scheduler.Task.run(Task.scala:109) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 18/11/12 11:27:42 ERROR org.apache.spark.scheduler.TaskSetManager: Task 46 in stage 0.0 failed 1 times; aborting job 18/11/12 11:27:42 WARN org.apache.spark.scheduler.TaskSetManager: Lost task 47.0 in stage 0.0 (TID 47, localhost, executor driver): java.io.FileNotFoundException: /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/1d/temp_shuffle_6bec34a8-e407-44d0-af66-4204fb9ea9b9 (No such file or directory) at java.io.FileOutputStream.open0(Native Method) at java.io.FileOutputStream.open(FileOutputStream.java:270) at java.io.FileOutputStream.<init>(FileOutputStream.java:213) at org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:103) at org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:116) at org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:237) at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:151) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) at org.apache.spark.scheduler.Task.run(Task.scala:109) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) 18/11/12 11:27:42 WARN org.apache.spark.storage.BlockManager: Putting block rdd_0_49 failed due to exception org.apache.spark.TaskKilledException. 18/11/12 11:27:42 WARN org.apache.spark.storage.BlockManager: Putting block rdd_0_48 failed due to exception org.apache.spark.TaskKilledException. 18/11/12 11:27:42 WARN org.apache.spark.storage.BlockManager: Block rdd_0_48 could not be removed as it was not found on disk or in memory 18/11/12 11:27:42 WARN org.apache.spark.storage.BlockManager: Block rdd_0_49 could not be removed as it was not found on disk or in memory 18/11/12 11:27:42 WARN org.apache.spark.storage.BlockManager: Putting block rdd_15_48 failed due to exception org.apache.spark.TaskKilledException. 18/11/12 11:27:42 WARN org.apache.spark.storage.BlockManager: Putting block rdd_15_49 failed due to exception org.apache.spark.TaskKilledException. 18/11/12 11:27:42 WARN org.apache.spark.storage.BlockManager: Block rdd_15_48 could not be removed as it was not found on disk or in memory 18/11/12 11:27:42 WARN org.apache.spark.storage.BlockManager: Block rdd_15_49 could not be removed as it was not found on disk or in memory 2018-11-12T11:27:42.539Z Running query:AVERAGE_SELLING_PRICE_BY_SELLER; exportSummaryToBigQuery:true; numEvents:10000; streamTimeout:60 2018-11-12T11:27:42.539Z skipping since configuration is not implemented 2018-11-12T11:27:42.539Z Running query:HIGHEST_BID; exportSummaryToBigQuery:true; streamTimeout:60 ========================================================================================== Run started 2018-11-12T11:27:04.746Z and ran for PT37.793S Default configuration: {"debug":true,"query":null,"sourceType":"DIRECT","sinkType":"DEVNULL","exportSummaryToBigQuery":false,"pubSubMode":"COMBINED","sideInputType":"DIRECT","sideInputRowCount":500,"sideInputNumShards":3,"sideInputUrl":null,"numEvents":100000,"numEventGenerators":100,"rateShape":"SINE","firstEventRate":10000,"nextEventRate":10000,"rateUnit":"PER_SECOND","ratePeriodSec":600,"preloadSeconds":0,"streamTimeout":240,"isRateLimited":false,"useWallclockEventTime":false,"avgPersonByteSize":200,"avgAuctionByteSize":500,"avgBidByteSize":100,"hotAuctionRatio":2,"hotSellersRatio":4,"hotBiddersRatio":4,"windowSizeSec":10,"windowPeriodSec":5,"watermarkHoldbackSec":0,"numInFlightAuctions":100,"numActivePeople":1000,"coderStrategy":"HAND","cpuDelayMs":0,"diskBusyBytes":0,"auctionSkip":123,"fanout":5,"maxAuctionsWaitingTime":600,"occasionalDelaySec":3,"probDelayedEvent":0.1,"maxLogEvents":100000,"usePubsubPublishTime":false,"outOfOrderGroupSize":1} Configurations: Conf Description 0000 query:PASSTHROUGH; exportSummaryToBigQuery:true; streamTimeout:60 0001 query:CURRENCY_CONVERSION; exportSummaryToBigQuery:true; streamTimeout:60 0002 query:SELECTION; exportSummaryToBigQuery:true; streamTimeout:60 0003 query:LOCAL_ITEM_SUGGESTION; exportSummaryToBigQuery:true; streamTimeout:60 Exception in thread "main" 0004 query:AVERAGE_PRICE_FOR_CATEGORY; exportSummaryToBigQuery:true; numEvents:10000; streamTimeout:60 0005 query:HOT_ITEMS; exportSummaryToBigQuery:true; streamTimeout:60 0006 query:AVERAGE_SELLING_PRICE_BY_SELLER; exportSummaryToBigQuery:true; numEvents:10000; streamTimeout:60 0007 query:HIGHEST_BID; exportSummaryToBigQuery:true; streamTimeout:60 0008 query:MONITOR_NEW_USERS; exportSummaryToBigQuery:true; streamTimeout:60 0009 query:WINNING_BIDS; exportSummaryToBigQuery:true; numEvents:10000; streamTimeout:60 0010 query:LOG_TO_SHARDED_FILES; exportSummaryToBigQuery:true; streamTimeout:60 java.lang.RuntimeException: org.apache.beam.sdk.Pipeline$PipelineExecutionException: java.io.FileNotFoundException: /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/3f/temp_shuffle_bc416c9b-bcb3-4c65-899a-f0df16c0fa6f (No such file or directory) at org.apache.beam.sdk.nexmark.Main.runAll(Main.java:147) at org.apache.beam.sdk.nexmark.Main.runAll(Main.java:116) at org.apache.beam.sdk.nexmark.Main.main(Main.java:480) 0011 query:USER_SESSIONS; exportSummaryToBigQuery:true; streamTimeout:60 0012 query:PROCESSING_TIME_WINDOWS; exportSummaryToBigQuery:true; streamTimeout:60 0013 query:BOUNDED_SIDE_INPUT_JOIN; exportSummaryToBigQuery:true; streamTimeout:60 Performance: Conf Runtime(sec) (Baseline) Events(/sec) (Baseline) Results (Baseline) 0000 2.7 37622.3 92000 0001 1.7 59594.8 92000 0002 1.5 68681.3 351 0003 5.6 17799.9 580 0004 3.5 2826.5 4 0005 *** not run *** 0006 *** not run *** 0007 *** not run *** 0008 *** not run *** 0009 *** not run *** 0010 *** not run *** 0011 *** not run *** 0012 *** not run *** 0013 *** not run *** ========================================================================================== 2018-11-12T11:27:42.540Z skipping since configuration is not implemented 2018-11-12T11:27:42.540Z Running query:MONITOR_NEW_USERS; exportSummaryToBigQuery:true; streamTimeout:60 2018-11-12T11:27:42.540Z skipping since configuration is not implemented 2018-11-12T11:27:42.540Z Running query:WINNING_BIDS; exportSummaryToBigQuery:true; numEvents:10000; streamTimeout:60 Caused by: org.apache.beam.sdk.Pipeline$PipelineExecutionException: java.io.FileNotFoundException: /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/3f/temp_shuffle_bc416c9b-bcb3-4c65-899a-f0df16c0fa6f (No such file or directory) at org.apache.beam.runners.spark.SparkPipelineResult.beamExceptionFrom(SparkPipelineResult.java:67) at org.apache.beam.runners.spark.SparkPipelineResult.waitUntilFinish(SparkPipelineResult.java:98) at org.apache.beam.sdk.nexmark.NexmarkLauncher.run(NexmarkLauncher.java:1163) at org.apache.beam.sdk.nexmark.Main$Run.call(Main.java:107) at org.apache.beam.sdk.nexmark.Main$Run.call(Main.java:96) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: java.io.FileNotFoundException: /tmp/blockmgr-96ecb829-7a5d-494d-b0af-f72f707dea23/3f/temp_shuffle_bc416c9b-bcb3-4c65-899a-f0df16c0fa6f (No such file or directory) at java.io.FileOutputStream.open0(Native Method) 2018-11-12T11:27:42.542Z at java.io.FileOutputStream.open(FileOutputStream.java:270) at java.io.FileOutputStream.<init>(FileOutputStream.java:213) skipping since configuration is not implemented at org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:103) 2018-11-12T11:27:42.542Z Running query:LOG_TO_SHARDED_FILES; exportSummaryToBigQuery:true; streamTimeout:60 at org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:116) at org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:237) at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:151) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96) at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) 2018-11-12T11:27:42.542Z skipping since configuration is not implemented at org.apache.spark.scheduler.Task.run(Task.scala:109) 2018-11-12T11:27:42.542Z Running query:USER_SESSIONS; exportSummaryToBigQuery:true; streamTimeout:60 at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345) ... 3 more 2018-11-12T11:27:42.543Z skipping since configuration is not implemented 2018-11-12T11:27:42.543Z Running query:PROCESSING_TIME_WINDOWS; exportSummaryToBigQuery:true; streamTimeout:60 2018-11-12T11:27:42.543Z skipping since configuration is not implemented 2018-11-12T11:27:42.543Z Running query:BOUNDED_SIDE_INPUT_JOIN; exportSummaryToBigQuery:true; streamTimeout:60 2018-11-12T11:27:42.543Z skipping since configuration is not implemented 18/11/12 11:27:42 ERROR org.apache.spark.util.Utils: Uncaught exception in thread Executor task launch worker for task 50 java.lang.NullPointerException at org.apache.spark.scheduler.Task$$anonfun$run$1.apply$mcV$sp(Task.scala:130) at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1361) at org.apache.spark.scheduler.Task.run(Task.scala:128) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) > Task :beam-sdks-java-nexmark:run FAILED :beam-sdks-java-nexmark:run (Thread[Daemon worker,5,main]) completed. Took 40.593 secs. FAILURE: Build failed with an exception. * What went wrong: Execution failed for task ':beam-sdks-java-nexmark:run'. > Process 'command '/usr/local/asfpackages/java/jdk1.8.0_191/bin/java'' > finished with non-zero exit value 1 * Try: Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights. * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 5.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/4.10.2/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 2m 13s 67 actionable tasks: 15 executed, 52 up-to-date Publishing build scan... https://gradle.com/s/dh3bee2hasfvy Build step 'Invoke Gradle script' changed build result to FAILURE Build step 'Invoke Gradle script' marked build as failure --------------------------------------------------------------------- To unsubscribe, e-mail: builds-unsubscr...@beam.apache.org For additional commands, e-mail: builds-h...@beam.apache.org