Do those also happen if you run other hadoop versions (e.g. try 1.0.4)?
On Tue, Jul 1, 2014 at 1:00 AM, Taka Shinagawa <taka.epsi...@gmail.com> wrote: > Since Spark 1.0.0, I've been seeing multiple errors when running sbt test. > > I ran the following commands from Spark 1.0.1 RC1 on Mac OSX 10.9.2. > > $ sbt/sbt clean > $ SPARK_HADOOP_VERSION=1.2.1 sbt/sbt assembly > $ sbt/sbt test > > > I'm attaching the log file generated by the sbt test. > > Here's the summary part of the test. > > [info] Run completed in 30 minutes, 57 seconds. > [info] Total number of tests run: 605 > [info] Suites: completed 83, aborted 0 > [info] Tests: succeeded 600, failed 5, canceled 0, ignored 5, pending 0 > [info] *** 5 TESTS FAILED *** > [error] Failed: Total 653, Failed 5, Errors 0, Passed 648, Ignored 5 > [error] Failed tests: > [error] org.apache.spark.ShuffleNettySuite > [error] org.apache.spark.ShuffleSuite > [error] org.apache.spark.FileServerSuite > [error] org.apache.spark.DistributedSuite > [error] (core/test:test) sbt.TestsFailedException: Tests unsuccessful > [error] Total time: 2033 s, completed Jul 1, 2014 12:08:03 AM > > Is anyone else seeing errors like this? > > > Thanks, > Taka