> Failed tests:
>   
> FrameMatrixReblockTest.testFrameWriteMultipleSparseBinarySpark:170->runFrameReblockTest:230
>  31 values are not in equal

What's going on with this test and the associated logic that it is testing?  
Isn't this the same test that has been intermittently failing for a while now?

-Mike

--

Mike Dusenberry
GitHub: github.com/dusenberrymw
LinkedIn: linkedin.com/in/mikedusenberry

Sent from my iPhone.


> On Feb 17, 2017, at 4:07 AM, jenk...@spark.tc wrote:
> 
> See <https://sparktc.ibmcloud.com/jenkins/job/SystemML-DailyTest/816/changes>
> 
> Changes:
> 
> [npansar] [MINOR] Code refactoring MatrixIndexingSPInstruction to enable 
> parallel
> 
> [Deron Eriksson] [SYSTEMML-1280] Restore and deprecate SQLContext methods
> 
> [Deron Eriksson] [SYSTEMML-1279] Decrease numCols to prevent spark codegen 
> issue
> 
> [Deron Eriksson] [SYSTEMML-1211] Update dependencies for Spark 2.1.0
> 
> [Deron Eriksson] [SYSTEMML-1271] Increment minimum Spark version in MLContext
> 
> [Deron Eriksson] [SYSTEMML-1277] MLContext implicitly convert mllib Vector to 
> ml Vector
> 
> [npansar] [MINOR] Addressed corner cases (i.e. empty blocks and extremely 
> large
> 
> ------------------------------------------
> [...truncated 12735 lines...]
> 17/02/17 04:59:57 INFO scheduler.DAGScheduler: Parents of final stage: 
> List(ShuffleMapStage 258)
> 17/02/17 04:59:57 INFO scheduler.DAGScheduler: Missing parents: 
> List(ShuffleMapStage 258)
> 17/02/17 04:59:57 INFO scheduler.DAGScheduler: Submitting ShuffleMapStage 258 
> (MapPartitionsRDD[715] at mapPartitionsToPair at RDDConverterUtils.java:256), 
> which has no missing parents
> 17/02/17 04:59:57 INFO memory.MemoryStore: Block broadcast_258 stored as 
> values in memory (estimated size 21.5 KB, free 1044.7 MB)
> 17/02/17 04:59:57 INFO memory.MemoryStore: Block broadcast_258_piece0 stored 
> as bytes in memory (estimated size 9.6 KB, free 1044.7 MB)
> 17/02/17 04:59:57 INFO storage.BlockManagerInfo: Added broadcast_258_piece0 
> in memory on 169.54.146.43:43583 (size: 9.6 KB, free: 1045.1 MB)
> 17/02/17 04:59:57 INFO spark.SparkContext: Created broadcast 258 from 
> broadcast at DAGScheduler.scala:996
> 17/02/17 04:59:57 INFO scheduler.DAGScheduler: Submitting 1 missing tasks 
> from ShuffleMapStage 258 (MapPartitionsRDD[715] at mapPartitionsToPair at 
> RDDConverterUtils.java:256)
> 17/02/17 04:59:57 INFO scheduler.TaskSchedulerImpl: Adding task set 258.0 
> with 1 tasks
> 17/02/17 04:59:57 INFO scheduler.FairSchedulableBuilder: Added task set 
> TaskSet_258.0 tasks to pool default
> 17/02/17 04:59:57 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 
> 258.0 (TID 268, localhost, executor driver, partition 0, PROCESS_LOCAL, 6289 
> bytes)
> 17/02/17 04:59:57 INFO executor.Executor: Running task 0.0 in stage 258.0 
> (TID 268)
> 17/02/17 04:59:57 INFO executor.Executor: Finished task 0.0 in stage 258.0 
> (TID 268). 1891 bytes result sent to driver
> 17/02/17 04:59:57 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 
> 258.0 (TID 268) in 13 ms on localhost (executor driver) (1/1)
> 17/02/17 04:59:57 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 258.0, 
> whose tasks have all completed, from pool default
> 17/02/17 04:59:57 INFO scheduler.DAGScheduler: ShuffleMapStage 258 
> (mapPartitionsToPair at RDDConverterUtils.java:256) finished in 0.015 s
> 17/02/17 04:59:57 INFO scheduler.DAGScheduler: looking for newly runnable 
> stages
> 17/02/17 04:59:57 INFO scheduler.DAGScheduler: running: Set()
> 17/02/17 04:59:57 INFO scheduler.DAGScheduler: waiting: Set(ResultStage 259)
> 17/02/17 04:59:57 INFO scheduler.DAGScheduler: failed: Set()
> 17/02/17 04:59:57 INFO scheduler.DAGScheduler: Submitting ResultStage 259 
> (MapPartitionsRDD[717] at mapValues at SparkUtils.java:147), which has no 
> missing parents
> 17/02/17 04:59:57 INFO memory.MemoryStore: Block broadcast_259 stored as 
> values in memory (estimated size 3.8 KB, free 1044.7 MB)
> 17/02/17 04:59:57 INFO memory.MemoryStore: Block broadcast_259_piece0 stored 
> as bytes in memory (estimated size 2.1 KB, free 1044.7 MB)
> 17/02/17 04:59:57 INFO storage.BlockManagerInfo: Added broadcast_259_piece0 
> in memory on 169.54.146.43:43583 (size: 2.1 KB, free: 1045.1 MB)
> 17/02/17 04:59:57 INFO spark.SparkContext: Created broadcast 259 from 
> broadcast at DAGScheduler.scala:996
> 17/02/17 04:59:57 INFO scheduler.DAGScheduler: Submitting 1 missing tasks 
> from ResultStage 259 (MapPartitionsRDD[717] at mapValues at 
> SparkUtils.java:147)
> 17/02/17 04:59:57 INFO scheduler.TaskSchedulerImpl: Adding task set 259.0 
> with 1 tasks
> 17/02/17 04:59:57 INFO scheduler.FairSchedulableBuilder: Added task set 
> TaskSet_259.0 tasks to pool default
> 17/02/17 04:59:57 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 
> 259.0 (TID 269, localhost, executor driver, partition 0, ANY, 5753 bytes)
> 17/02/17 04:59:57 INFO executor.Executor: Running task 0.0 in stage 259.0 
> (TID 269)
> 17/02/17 04:59:57 INFO storage.ShuffleBlockFetcherIterator: Getting 1 
> non-empty blocks out of 1 blocks
> 17/02/17 04:59:57 INFO storage.ShuffleBlockFetcherIterator: Started 0 remote 
> fetches in 0 ms
> 17/02/17 04:59:57 INFO executor.Executor: Finished task 0.0 in stage 259.0 
> (TID 269). 2021 bytes result sent to driver
> 17/02/17 04:59:57 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 
> 259.0 (TID 269) in 5 ms on localhost (executor driver) (1/1)
> 17/02/17 04:59:57 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 259.0, 
> whose tasks have all completed, from pool default
> 17/02/17 04:59:57 INFO scheduler.DAGScheduler: ResultStage 259 (collect at 
> SparkExecutionContext.java:768) finished in 0.006 s
> 17/02/17 04:59:58 INFO storage.BlockManagerInfo: Removed broadcast_258_piece0 
> on 169.54.146.43:43583 in memory (size: 9.6 KB, free: 1045.1 MB)
> 17/02/17 04:59:58 INFO spark.ContextCleaner: Cleaned accumulator 16831
> 17/02/17 04:59:58 INFO storage.BlockManagerInfo: Removed broadcast_259_piece0 
> on 169.54.146.43:43583 in memory (size: 2.1 KB, free: 1045.1 MB)
> 17/02/17 04:59:58 INFO storage.BlockManagerInfo: Removed broadcast_257_piece0 
> on 169.54.146.43:43583 in memory (size: 9.0 KB, free: 1045.2 MB)
> 17/02/17 04:59:58 INFO storage.BlockManagerInfo: Removed broadcast_256_piece0 
> on 169.54.146.43:43583 in memory (size: 8.9 KB, free: 1045.2 MB)
> 17/02/17 04:59:58 INFO scheduler.DAGScheduler: Registering RDD 719 
> (flatMapToPair at RDDConverterUtils.java:273)
> 17/02/17 04:59:58 INFO scheduler.DAGScheduler: Got job 193 (collectAsList at 
> MLContextTest.java:1343) with 1 output partitions
> 17/02/17 04:59:58 INFO scheduler.DAGScheduler: Final stage: ResultStage 261 
> (collectAsList at MLContextTest.java:1343)
> 17/02/17 04:59:58 INFO scheduler.DAGScheduler: Parents of final stage: 
> List(ShuffleMapStage 260)
> 17/02/17 04:59:58 INFO scheduler.DAGScheduler: Missing parents: 
> List(ShuffleMapStage 260)
> 17/02/17 04:59:58 INFO scheduler.DAGScheduler: Submitting ShuffleMapStage 260 
> (MapPartitionsRDD[719] at flatMapToPair at RDDConverterUtils.java:273), which 
> has no missing parents
> 17/02/17 04:59:58 INFO memory.MemoryStore: Block broadcast_260 stored as 
> values in memory (estimated size 3.9 KB, free 1044.7 MB)
> 17/02/17 04:59:58 INFO memory.MemoryStore: Block broadcast_260_piece0 stored 
> as bytes in memory (estimated size 2.2 KB, free 1044.7 MB)
> 17/02/17 04:59:58 INFO storage.BlockManagerInfo: Added broadcast_260_piece0 
> in memory on 169.54.146.43:43583 (size: 2.2 KB, free: 1045.2 MB)
> 17/02/17 04:59:58 INFO spark.SparkContext: Created broadcast 260 from 
> broadcast at DAGScheduler.scala:996
> 17/02/17 04:59:58 INFO scheduler.DAGScheduler: Submitting 1 missing tasks 
> from ShuffleMapStage 260 (MapPartitionsRDD[719] at flatMapToPair at 
> RDDConverterUtils.java:273)
> 17/02/17 04:59:58 INFO scheduler.TaskSchedulerImpl: Adding task set 260.0 
> with 1 tasks
> 17/02/17 04:59:58 INFO scheduler.FairSchedulableBuilder: Added task set 
> TaskSet_260.0 tasks to pool default
> 17/02/17 04:59:58 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 
> 260.0 (TID 270, localhost, executor driver, partition 0, PROCESS_LOCAL, 6218 
> bytes)
> 17/02/17 04:59:58 INFO executor.Executor: Running task 0.0 in stage 260.0 
> (TID 270)
> 17/02/17 04:59:58 INFO executor.Executor: Finished task 0.0 in stage 260.0 
> (TID 270). 1253 bytes result sent to driver
> 17/02/17 04:59:58 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 
> 260.0 (TID 270) in 6 ms on localhost (executor driver) (1/1)
> 17/02/17 04:59:58 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 260.0, 
> whose tasks have all completed, from pool default
> 17/02/17 04:59:58 INFO scheduler.DAGScheduler: ShuffleMapStage 260 
> (flatMapToPair at RDDConverterUtils.java:273) finished in 0.006 s
> 17/02/17 04:59:58 INFO scheduler.DAGScheduler: looking for newly runnable 
> stages
> 17/02/17 04:59:58 INFO scheduler.DAGScheduler: running: Set()
> 17/02/17 04:59:58 INFO scheduler.DAGScheduler: waiting: Set(ResultStage 261)
> 17/02/17 04:59:58 INFO scheduler.DAGScheduler: failed: Set()
> 17/02/17 04:59:58 INFO scheduler.DAGScheduler: Submitting ResultStage 261 
> (MapPartitionsRDD[725] at collectAsList at MLContextTest.java:1343), which 
> has no missing parents
> 17/02/17 04:59:58 INFO memory.MemoryStore: Block broadcast_261 stored as 
> values in memory (estimated size 11.3 KB, free 1044.7 MB)
> 17/02/17 04:59:58 INFO memory.MemoryStore: Block broadcast_261_piece0 stored 
> as bytes in memory (estimated size 5.6 KB, free 1044.7 MB)
> 17/02/17 04:59:58 INFO storage.BlockManagerInfo: Added broadcast_261_piece0 
> in memory on 169.54.146.43:43583 (size: 5.6 KB, free: 1045.2 MB)
> 17/02/17 04:59:58 INFO spark.SparkContext: Created broadcast 261 from 
> broadcast at DAGScheduler.scala:996
> 17/02/17 04:59:58 INFO scheduler.DAGScheduler: Submitting 1 missing tasks 
> from ResultStage 261 (MapPartitionsRDD[725] at collectAsList at 
> MLContextTest.java:1343)
> 17/02/17 04:59:58 INFO scheduler.TaskSchedulerImpl: Adding task set 261.0 
> with 1 tasks
> 17/02/17 04:59:58 INFO scheduler.FairSchedulableBuilder: Added task set 
> TaskSet_261.0 tasks to pool default
> 17/02/17 04:59:58 INFO scheduler.TaskSetManager: Starting task 0.0 in stage 
> 261.0 (TID 271, localhost, executor driver, partition 0, ANY, 5783 bytes)
> 17/02/17 04:59:58 INFO executor.Executor: Running task 0.0 in stage 261.0 
> (TID 271)
> 17/02/17 04:59:58 INFO storage.ShuffleBlockFetcherIterator: Getting 1 
> non-empty blocks out of 1 blocks
> 17/02/17 04:59:58 INFO storage.ShuffleBlockFetcherIterator: Started 0 remote 
> fetches in 1 ms
> 17/02/17 04:59:58 INFO executor.Executor: Finished task 0.0 in stage 261.0 
> (TID 271). 2024 bytes result sent to driver
> 17/02/17 04:59:58 INFO scheduler.TaskSetManager: Finished task 0.0 in stage 
> 261.0 (TID 271) in 7 ms on localhost (executor driver) (1/1)
> 17/02/17 04:59:58 INFO scheduler.TaskSchedulerImpl: Removed TaskSet 261.0, 
> whose tasks have all completed, from pool default
> 17/02/17 04:59:58 INFO scheduler.DAGScheduler: ResultStage 261 (collectAsList 
> at MLContextTest.java:1343) finished in 0.009 s
> 17/02/17 04:59:59 INFO storage.BlockManagerInfo: Removed broadcast_261_piece0 
> on 169.54.146.43:43583 in memory (size: 5.6 KB, free: 1045.2 MB)
> 17/02/17 04:59:59 INFO storage.BlockManagerInfo: Removed broadcast_260_piece0 
> on 169.54.146.43:43583 in memory (size: 2.2 KB, free: 1045.2 MB)
> 17/02/17 04:59:59 INFO spark.ContextCleaner: Cleaned accumulator 17026
> 17/02/17 04:59:59 INFO spark.ContextCleaner: Cleaned shuffle 67
> 17/02/17 04:59:59 INFO server.ServerConnector: Stopped 
> ServerConnector@2aaced77{HTTP/1.1}{0.0.0.0:4040}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@1deddf65{/stages/stage/kill,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@70677659{/jobs/job/kill,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@4aabd893{/api,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@7529412e{/,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@6be0d0e0{/static,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@5f7ca14e{/executors/threadDump/json,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@7a68aaa2{/executors/threadDump,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@5e4813e9{/executors/json,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@7278ca13{/executors,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@69998504{/environment/json,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@283196b4{/environment,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@15c73b7e{/storage/rdd/json,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@693e589c{/storage/rdd,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@5c53b749{/storage/json,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@636bd416{/storage,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@38588bbe{/stages/pool/json,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@4cff34f3{/stages/pool,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@364138da{/stages/stage/json,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@3e2ca893{/stages/stage,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@223acd27{/stages/json,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@53f5263f{/stages,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@1cf282fe{/jobs/job/json,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@2956ab98{/jobs/job,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@6d0387a9{/jobs/json,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO handler.ContextHandler: Stopped 
> o.s.j.s.ServletContextHandler@40b1ef30{/jobs,null,UNAVAILABLE}
> 17/02/17 04:59:59 INFO ui.SparkUI: Stopped Spark web UI at 
> http://169.54.146.43:4040
> 17/02/17 04:59:59 INFO spark.MapOutputTrackerMasterEndpoint: 
> MapOutputTrackerMasterEndpoint stopped!
> 17/02/17 04:59:59 INFO memory.MemoryStore: MemoryStore cleared
> 17/02/17 04:59:59 INFO storage.BlockManager: BlockManager stopped
> 17/02/17 04:59:59 INFO storage.BlockManagerMaster: BlockManagerMaster stopped
> 17/02/17 04:59:59 INFO 
> scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: 
> OutputCommitCoordinator stopped!
> 17/02/17 04:59:59 INFO spark.SparkContext: Successfully stopped SparkContext
> Running org.apache.sysml.test.integration.mlcontext.MLContextTest
> Tests run: 169, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 63.734 sec 
> - in org.apache.sysml.test.integration.mlcontext.MLContextTest
> 17/02/17 04:59:59 INFO util.ShutdownHookManager: Shutdown hook called
> 17/02/17 04:59:59 INFO util.ShutdownHookManager: Deleting directory 
> /tmp/spark-4b81537e-e1c4-4b89-8821-3efd1c0dc2d8
> Running org.apache.sysml.test.integration.functions.indexing.LeftIndexingTest
> Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 678.607 sec - 
> in org.apache.sysml.test.integration.functions.indexing.LeftIndexingTest
> Running org.apache.sysml.test.integration.functions.append.AppendVectorTest
> Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 432.935 sec - 
> in org.apache.sysml.test.integration.functions.append.AppendVectorTest
> Running 
> org.apache.sysml.test.integration.functions.quaternary.WeightedSigmoidTest
> Tests run: 56, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1,004.868 
> sec - in 
> org.apache.sysml.test.integration.functions.quaternary.WeightedSigmoidTest
> 17/02/17 05:02:56 INFO util.ShutdownHookManager: Shutdown hook called
> 17/02/17 05:02:56 INFO util.ShutdownHookManager: Deleting directory 
> /tmp/spark-bf095cc1-cf00-4bfd-8a06-7c279a743c14
> Running 
> org.apache.sysml.test.integration.applications.parfor.ParForCorrelationTest
> Tests run: 12, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 285.721 sec 
> - in 
> org.apache.sysml.test.integration.applications.parfor.ParForCorrelationTest
> Running 
> org.apache.sysml.test.integration.applications.descriptivestats.UnivariateUnweightedScaleDenseTest
> Tests run: 24, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 348.853 sec 
> - in 
> org.apache.sysml.test.integration.applications.descriptivestats.UnivariateUnweightedScaleDenseTest
> Running 
> org.apache.sysml.test.integration.applications.parfor.ParForCorrelationTestLarge
> Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 43.071 sec - 
> in 
> org.apache.sysml.test.integration.applications.parfor.ParForCorrelationTestLarge
> Running org.apache.sysml.test.integration.functions.reorg.FullOrderTest
> Tests run: 132, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1,368.647 
> sec - in org.apache.sysml.test.integration.functions.reorg.FullOrderTest
> Running 
> org.apache.sysml.test.integration.functions.quaternary.WeightedDivMatrixMultTest
> Tests run: 90, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1,489.678 
> sec - in 
> org.apache.sysml.test.integration.functions.quaternary.WeightedDivMatrixMultTest
> 17/02/17 05:09:08 INFO util.ShutdownHookManager: Shutdown hook called
> 17/02/17 05:09:08 INFO util.ShutdownHookManager: Deleting directory 
> /tmp/spark-fb94f0d9-b2b8-4ab3-bd22-59162e93e33e
> Running 
> org.apache.sysml.test.integration.applications.descriptivestats.UnivariateWeightedScaleDenseTest
> Tests run: 24, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 345.726 sec 
> - in 
> org.apache.sysml.test.integration.applications.descriptivestats.UnivariateWeightedScaleDenseTest
> Running 
> org.apache.sysml.test.integration.applications.parfor.ParForNaiveBayesTest
> Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 531.889 sec 
> - in 
> org.apache.sysml.test.integration.applications.parfor.ParForNaiveBayesTest
> Running org.apache.sysml.test.integration.functions.append.AppendChainTest
> Tests run: 12, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1,062.068 
> sec - in org.apache.sysml.test.integration.functions.append.AppendChainTest
> Running 
> org.apache.sysml.test.integration.applications.descriptivestats.UnivariateUnweightedScaleSparseTest
> Tests run: 24, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 277.404 sec 
> - in 
> org.apache.sysml.test.integration.applications.descriptivestats.UnivariateUnweightedScaleSparseTest
> Running org.apache.sysml.test.integration.applications.dml.HITSDMLTest
> Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 92.16 sec - 
> in org.apache.sysml.test.integration.applications.dml.HITSDMLTest
> Running org.apache.sysml.test.integration.functions.append.AppendMatrixTest
> Tests run: 22, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1,430.611 
> sec - in org.apache.sysml.test.integration.functions.append.AppendMatrixTest
> Running 
> org.apache.sysml.test.integration.applications.descriptivestats.UnivariateWeightedScaleSparseTest
> Tests run: 24, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 329.608 sec 
> - in 
> org.apache.sysml.test.integration.applications.descriptivestats.UnivariateWeightedScaleSparseTest
> Running org.apache.sysml.test.integration.applications.pydml.ArimaPyDMLTest
> Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 49.175 sec - 
> in org.apache.sysml.test.integration.applications.pydml.ArimaPyDMLTest
> Running org.apache.sysml.test.integration.applications.pydml.HITSPyDMLTest
> Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 76.627 sec - 
> in org.apache.sysml.test.integration.applications.pydml.HITSPyDMLTest
> Running 
> org.apache.sysml.test.integration.applications.dml.LinearRegressionDMLTest
> Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 51.395 sec - 
> in org.apache.sysml.test.integration.applications.dml.LinearRegressionDMLTest
> Running 
> org.apache.sysml.test.integration.applications.pydml.CsplineDSPyDMLTest
> Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.831 sec - 
> in org.apache.sysml.test.integration.applications.pydml.CsplineDSPyDMLTest
> Running 
> org.apache.sysml.test.integration.applications.pydml.LinearLogRegPyDMLTest
> Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 82.606 sec - 
> in org.apache.sysml.test.integration.applications.pydml.LinearLogRegPyDMLTest
> Running 
> org.apache.sysml.test.integration.applications.pydml.NaiveBayesPyDMLTest
> Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 81.42 sec - 
> in org.apache.sysml.test.integration.applications.pydml.NaiveBayesPyDMLTest
> Running org.apache.sysml.test.integration.applications.dml.CsplineCGDMLTest
> Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 15.806 sec - 
> in org.apache.sysml.test.integration.applications.dml.CsplineCGDMLTest
> Running org.apache.sysml.test.integration.applications.dml.L2SVMDMLTest
> Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 64.378 sec - 
> in org.apache.sysml.test.integration.applications.dml.L2SVMDMLTest
> Running 
> org.apache.sysml.test.integration.applications.pydml.MDABivariateStatsPyDMLTest
> Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 294.972 sec - 
> in 
> org.apache.sysml.test.integration.applications.pydml.MDABivariateStatsPyDMLTest
> Running org.apache.sysml.test.integration.applications.pydml.PageRankPyDMLTest
> Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.3 sec - in 
> org.apache.sysml.test.integration.applications.pydml.PageRankPyDMLTest
> Running 
> org.apache.sysml.test.integration.applications.pydml.NaiveBayesParforPyDMLTest
> Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 81.195 sec - 
> in 
> org.apache.sysml.test.integration.applications.pydml.NaiveBayesParforPyDMLTest
> Running org.apache.sysml.test.integration.applications.dml.NaiveBayesDMLTest
> Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 81.784 sec - 
> in org.apache.sysml.test.integration.applications.dml.NaiveBayesDMLTest
> Running 
> org.apache.sysml.test.integration.applications.dml.MDABivariateStatsDMLTest
> Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 293.84 sec - 
> in org.apache.sysml.test.integration.applications.dml.MDABivariateStatsDMLTest
> Running org.apache.sysml.test.integration.applications.dml.PageRankDMLTest
> Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.301 sec - 
> in org.apache.sysml.test.integration.applications.dml.PageRankDMLTest
> Running org.apache.sysml.test.integration.applications.pydml.GLMPyDMLTest
> Tests run: 30, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 586.548 sec 
> - in org.apache.sysml.test.integration.applications.pydml.GLMPyDMLTest
> Running 
> org.apache.sysml.test.integration.applications.pydml.MultiClassSVMPyDMLTest
> Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 89.255 sec - 
> in org.apache.sysml.test.integration.applications.pydml.MultiClassSVMPyDMLTest
> Running org.apache.sysml.test.integration.applications.dml.ID3DMLTest
> Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 390.245 sec - 
> in org.apache.sysml.test.integration.applications.dml.ID3DMLTest
> Running org.apache.sysml.test.integration.applications.dml.GLMDMLTest
> Tests run: 30, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 889.89 sec - 
> in org.apache.sysml.test.integration.applications.dml.GLMDMLTest
> Running 
> org.apache.sysml.test.integration.applications.pydml.CsplineCGPyDMLTest
> Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.232 sec - 
> in org.apache.sysml.test.integration.applications.pydml.CsplineCGPyDMLTest
> Running org.apache.sysml.test.integration.applications.pydml.WelchTPyDMLTest
> Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.629 sec - 
> in org.apache.sysml.test.integration.applications.pydml.WelchTPyDMLTest
> Running 
> org.apache.sysml.test.integration.applications.dml.MultiClassSVMDMLTest
> Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 87.486 sec - 
> in org.apache.sysml.test.integration.applications.dml.MultiClassSVMDMLTest
> Running org.apache.sysml.test.integration.applications.dml.CsplineDSDMLTest
> Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 12.518 sec - 
> in org.apache.sysml.test.integration.applications.dml.CsplineDSDMLTest
> Running 
> org.apache.sysml.test.integration.applications.pydml.ApplyTransformPyDMLTest
> Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.26 sec - in 
> org.apache.sysml.test.integration.applications.pydml.ApplyTransformPyDMLTest
> Running 
> org.apache.sysml.test.integration.applications.dml.NaiveBayesParforDMLTest
> Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 81.105 sec - 
> in org.apache.sysml.test.integration.applications.dml.NaiveBayesParforDMLTest
> Running org.apache.sysml.test.integration.applications.pydml.L2SVMPyDMLTest
> Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 64.522 sec - 
> in org.apache.sysml.test.integration.applications.pydml.L2SVMPyDMLTest
> Running 
> org.apache.sysml.test.integration.applications.pydml.LinearRegressionPyDMLTest
> Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 51.556 sec - 
> in 
> org.apache.sysml.test.integration.applications.pydml.LinearRegressionPyDMLTest
> Running org.apache.sysml.test.integration.applications.dml.LinearLogRegDMLTest
> Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 61.237 sec - 
> in org.apache.sysml.test.integration.applications.dml.LinearLogRegDMLTest
> Running org.apache.sysml.test.integration.applications.dml.GNMFDMLTest
> Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 56.504 sec - 
> in org.apache.sysml.test.integration.applications.dml.GNMFDMLTest
> Running org.apache.sysml.test.integration.applications.dml.WelchTDMLTest
> Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.657 sec - 
> in org.apache.sysml.test.integration.applications.dml.WelchTDMLTest
> Running 
> org.apache.sysml.test.integration.applications.dml.ApplyTransformDMLTest
> Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 12.661 sec - 
> in org.apache.sysml.test.integration.applications.dml.ApplyTransformDMLTest
> Running org.apache.sysml.test.integration.applications.pydml.ID3PyDMLTest
> Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 379.455 sec - 
> in org.apache.sysml.test.integration.applications.pydml.ID3PyDMLTest
> Running org.apache.sysml.test.integration.applications.pydml.GNMFPyDMLTest
> Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 57.108 sec - 
> in org.apache.sysml.test.integration.applications.pydml.GNMFPyDMLTest
> 
> Results :
> 
> Failed tests: 
>  
> FrameMatrixReblockTest.testFrameWriteMultipleSparseBinarySpark:170->runFrameReblockTest:230
>  31 values are not in equal
> 
> Tests run: 6608, Failures: 1, Errors: 0, Skipped: 0
> 
> [INFO] 
> [INFO] --- maven-failsafe-plugin:2.17:verify (default) @ systemml ---
> [INFO] Failsafe report directory: 
> <https://sparktc.ibmcloud.com/jenkins/job/SystemML-DailyTest/ws/target/failsafe-reports>
> [INFO] 
> ------------------------------------------------------------------------
> [INFO] BUILD FAILURE
> [INFO] 
> ------------------------------------------------------------------------
> [INFO] Total time: 02:25 h
> [INFO] Finished at: 2017-02-17T06:07:07-06:00
> [INFO] Final Memory: 67M/2167M
> [INFO] 
> ------------------------------------------------------------------------
> [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-failsafe-plugin:2.17:verify (default) on 
> project systemml: There are test failures.
> [ERROR] 
> [ERROR] Please refer to 
> <https://sparktc.ibmcloud.com/jenkins/job/SystemML-DailyTest/ws/target/failsafe-reports>
>  for the individual test results.
> [ERROR] -> [Help 1]
> [ERROR] 
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR] 
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
> Build step 'Execute shell' marked build as failure
> Run condition [Always] enabling perform for step [[]]

Reply via email to