See <https://builds.apache.org/job/Mahout-Quality/2889/>

------------------------------------------
[...truncated 6221 lines...]
{
  0  => {0:0.5351226481634418,1:0.048656484626039664,2:0.7776206844268364}
  1  => {0:0.43076743320865907,1:0.5015041675073865,2:0.5638959606657767}
  2  => {0:0.08806393416547864,1:0.6374927759347693,2:0.1823970345980125}
}
C=
{
  0  => {0:1.535122648163442,1:2.0486564846260396,2:3.7776206844268363}
  1  => {0:3.430767433208659,1:4.501504167507386,2:5.563895960665777}
  2  => {0:5.088063934165478,1:6.637492775934769,2:7.182397034598012}
}
- C = A + B, identically partitioned
- C = A + B side test 1
- C = A + B side test 2
- C = A + B side test 3
- Ax
- A'x
- colSums, colMeans
- rowSums, rowMeans
- A.diagv
- numNonZeroElementsPerColumn
- C = A cbind B, cogroup
- C = A cbind B, zip
- B = A + 1.0
- C = A rbind B
- C = A rbind B, with empty
- scalarOps
1 [Executor task launch worker-0] ERROR org.apache.spark.executor.Executor  - 
Exception in task 4.0 in stage 245.0 (TID 538)
java.io.IOException: PARSING_ERROR(2)
        at org.xerial.snappy.SnappyNative.throw_error(SnappyNative.java:78)
        at org.xerial.snappy.SnappyNative.uncompressedLength(Native Method)
        at org.xerial.snappy.Snappy.uncompressedLength(Snappy.java:545)
        at 
org.xerial.snappy.SnappyInputStream.readFully(SnappyInputStream.java:125)
        at 
org.xerial.snappy.SnappyInputStream.readHeader(SnappyInputStream.java:88)
        at org.xerial.snappy.SnappyInputStream.<init>(SnappyInputStream.java:58)
        at 
org.apache.spark.io.SnappyCompressionCodec.compressedInputStream(CompressionCodec.scala:128)
        at 
org.apache.spark.broadcast.TorrentBroadcast$.unBlockifyObject(TorrentBroadcast.scala:232)
        at 
org.apache.spark.broadcast.TorrentBroadcast.readObject(TorrentBroadcast.scala:169)
        at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at 
java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:969)
        at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1871)
        at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1775)
        at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1327)
        at 
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1969)
        at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893)
        at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1775)
        at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1327)
        at java.io.ObjectInputStream.readObject(ObjectInputStream.java:349)
        at 
org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:62)
        at 
org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:87)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:159)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
        at java.lang.Thread.run(Thread.java:662)
1 [Executor task launch worker-2] ERROR org.apache.spark.executor.Executor  - 
Exception in task 5.0 in stage 245.0 (TID 539)
java.io.IOException: PARSING_ERROR(2)
        at org.xerial.snappy.SnappyNative.throw_error(SnappyNative.java:78)
        at org.xerial.snappy.SnappyNative.uncompressedLength(Native Method)
        at org.xerial.snappy.Snappy.uncompressedLength(Snappy.java:545)
        at 
org.xerial.snappy.SnappyInputStream.readFully(SnappyInputStream.java:125)
        at 
org.xerial.snappy.SnappyInputStream.readHeader(SnappyInputStream.java:88)
        at org.xerial.snappy.SnappyInputStream.<init>(SnappyInputStream.java:58)
        at 
org.apache.spark.io.SnappyCompressionCodec.compressedInputStream(CompressionCodec.scala:128)
        at 
org.apache.spark.broadcast.TorrentBroadcast$.unBlockifyObject(TorrentBroadcast.scala:232)
        at 
org.apache.spark.broadcast.TorrentBroadcast.readObject(TorrentBroadcast.scala:169)
        at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at 
java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:969)
        at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1871)
        at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1775)
        at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1327)
        at 
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1969)
        at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893)
        at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1775)
        at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1327)
        at java.io.ObjectInputStream.readObject(ObjectInputStream.java:349)
        at 
org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:62)
        at 
org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:87)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:159)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
        at java.lang.Thread.run(Thread.java:662)
1 [Executor task launch worker-1] ERROR org.apache.spark.executor.Executor  - 
Exception in task 6.0 in stage 245.0 (TID 540)
java.io.IOException: PARSING_ERROR(2)
        at org.xerial.snappy.SnappyNative.throw_error(SnappyNative.java:78)
        at org.xerial.snappy.SnappyNative.uncompressedLength(Native Method)
        at org.xerial.snappy.Snappy.uncompressedLength(Snappy.java:545)
        at 
org.xerial.snappy.SnappyInputStream.readFully(SnappyInputStream.java:125)
        at 
org.xerial.snappy.SnappyInputStream.readHeader(SnappyInputStream.java:88)
        at org.xerial.snappy.SnappyInputStream.<init>(SnappyInputStream.java:58)
        at 
org.apache.spark.io.SnappyCompressionCodec.compressedInputStream(CompressionCodec.scala:128)
        at 
org.apache.spark.broadcast.TorrentBroadcast$.unBlockifyObject(TorrentBroadcast.scala:232)
        at 
org.apache.spark.broadcast.TorrentBroadcast.readObject(TorrentBroadcast.scala:169)
        at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at 
java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:969)
        at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1871)
        at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1775)
        at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1327)
        at 
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1969)
        at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893)
        at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1775)
        at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1327)
        at java.io.ObjectInputStream.readObject(ObjectInputStream.java:349)
        at 
org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:62)
        at 
org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:87)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:159)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
        at java.lang.Thread.run(Thread.java:662)
13 [Executor task launch worker-2] ERROR org.apache.spark.executor.Executor  - 
Exception in task 7.0 in stage 245.0 (TID 541)
java.io.IOException: PARSING_ERROR(2)
        at org.xerial.snappy.SnappyNative.throw_error(SnappyNative.java:78)
        at org.xerial.snappy.SnappyNative.uncompressedLength(Native Method)
        at org.xerial.snappy.Snappy.uncompressedLength(Snappy.java:545)
        at 
org.xerial.snappy.SnappyInputStream.readFully(SnappyInputStream.java:125)
        at 
org.xerial.snappy.SnappyInputStream.readHeader(SnappyInputStream.java:88)
        at org.xerial.snappy.SnappyInputStream.<init>(SnappyInputStream.java:58)
        at 
org.apache.spark.io.SnappyCompressionCodec.compressedInputStream(CompressionCodec.scala:128)
        at 
org.apache.spark.broadcast.TorrentBroadcast$.unBlockifyObject(TorrentBroadcast.scala:232)
        at 
org.apache.spark.broadcast.TorrentBroadcast.readObject(TorrentBroadcast.scala:169)
        at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at 
java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:969)
        at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1871)
        at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1775)
        at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1327)
        at 
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1969)
        at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893)
        at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1775)
        at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1327)
        at java.io.ObjectInputStream.readObject(ObjectInputStream.java:349)
        at 
org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:62)
        at 
org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:87)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:159)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
        at java.lang.Thread.run(Thread.java:662)
15 [Result resolver thread-2] ERROR org.apache.spark.scheduler.TaskSetManager  
- Task 4 in stage 245.0 failed 1 times; aborting job
- C = A + B missing rows *** FAILED ***
  org.apache.spark.SparkException: Job aborted due to stage failure: Task 
4 in stage 245.0 failed 1 times, most recent failure: Lost task 4.0 in stage 
245.0 (TID 538, localhost): java.io.IOException: PARSING_ERROR(2)
        
org.xerial.snappy.SnappyNative.throw_error(SnappyNative.java:78)
        org.xerial.snappy.SnappyNative.uncompressedLength(Native 
Method)
        org.xerial.snappy.Snappy.uncompressedLength(Snappy.java:545)
        
org.xerial.snappy.SnappyInputStream.readFully(SnappyInputStream.java:125)
        
org.xerial.snappy.SnappyInputStream.readHeader(SnappyInputStream.java:88)
        
org.xerial.snappy.SnappyInputStream.<init>(SnappyInputStream.java:58)
        
org.apache.spark.io.SnappyCompressionCodec.compressedInputStream(CompressionCodec.scala:128)
        
org.apache.spark.broadcast.TorrentBroadcast$.unBlockifyObject(TorrentBroadcast.scala:232)
        
org.apache.spark.broadcast.TorrentBroadcast.readObject(TorrentBroadcast.scala:169)
        sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
        
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        java.lang.reflect.Method.invoke(Method.java:597)
        
java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:969)
        
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1871)
        
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1775)
        
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1327)
        
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1969)
        
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893)
        
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1775)
        
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1327)
        
java.io.ObjectInputStream.readObject(ObjectInputStream.java:349)
        
org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:62)
        
org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:87)
        
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:159)
        
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895)
        
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918)
        java.lang.Thread.run(Thread.java:662)
Driver stacktrace:
  at 
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1185)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1174)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1173)
  at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
  at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
  at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1173)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:688)
  at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:688)
  at scala.Option.foreach(Option.scala:236)
  at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:688)
  ...
- C = cbind(A, B) with missing rows
collected A = 
{
  0  => {0:1.0,1:2.0,2:3.0}
  1  => {}
  2  => {}
  3  => {0:3.0,1:4.0,2:5.0}
}
collected B = 
{
  2  => {0:1.0,1:1.0,2:1.0}
  1  => {0:1.0,1:1.0,2:1.0}
  3  => {0:4.0,1:5.0,2:6.0}
  0  => {0:2.0,1:3.0,2:4.0}
}
- B = A + 1.0 missing rows
Run completed in 2 minutes, 46 seconds.
Total number of tests run: 75
Suites: completed 10, aborted 0
Tests: succeeded 74, failed 1, canceled 0, ignored 1, pending 0
*** 1 TEST FAILED ***
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Mahout Build Tools ................................ SUCCESS [4.610s]
[INFO] Apache Mahout ..................................... SUCCESS [1.922s]
[INFO] Mahout Math ....................................... SUCCESS [2:22.833s]
[INFO] Mahout MapReduce Legacy ........................... SUCCESS [11:47.213s]
[INFO] Mahout Integration ................................ SUCCESS [1:26.887s]
[INFO] Mahout Examples ................................... SUCCESS [52.782s]
[INFO] Mahout Release Package ............................ SUCCESS [0.114s]
[INFO] Mahout Math Scala bindings ........................ SUCCESS [2:09.594s]
[INFO] Mahout Spark bindings ............................. FAILURE [3:29.014s]
[INFO] Mahout Spark bindings shell ....................... SKIPPED
[INFO] Mahout H2O backend ................................ SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 22:23.265s
[INFO] Finished at: Wed Dec 10 17:36:39 UTC 2014
[INFO] Final Memory: 88M/430M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.scalatest:scalatest-maven-plugin:1.0-M2:test 
(test) on project mahout-spark_2.10: There are test failures -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :mahout-spark_2.10
Build step 'Invoke top-level Maven targets' marked build as failure
[PMD] Skipping publisher since build result is FAILURE
[TASKS] Skipping publisher since build result is FAILURE
Archiving artifacts
Sending artifact delta relative to Mahout-Quality #2888
Archived 72 artifacts
Archive block size is 32768
Received 3624 blocks and 20625150 bytes
Compression is 85.2%
Took 12 sec
Recording test results
Publishing Javadoc

Reply via email to