Yes.
That could be the cause.

On Sun, Jan 18, 2015 at 11:47 AM, Sean Owen <so...@cloudera.com> wrote:

> Oh: are you running the tests with a different profile setting than
> what the last assembly was built with? this particular test depends on
> those matching. Not 100% sure that's the problem, but a good guess.
>
> On Sat, Jan 17, 2015 at 4:54 PM, Ted Yu <yuzhih...@gmail.com> wrote:
> > The test passed here:
> >
> >
> https://amplab.cs.berkeley.edu/jenkins/view/Spark/job/Spark-Master-Maven-with-YARN/HADOOP_PROFILE=hadoop-2.4,label=centos/1215/consoleFull
> >
> > It passed locally with the following command:
> >
> > mvn -DHADOOP_PROFILE=hadoop-2.4 -Phadoop-2.4 -Pyarn -Phive test
> > -Dtest=JavaAPISuite
> >
> > FYI
> >
> > On Sat, Jan 17, 2015 at 8:23 AM, Andrew Musselman
> > <andrew.mussel...@gmail.com> wrote:
> >>
> >> Failing for me and another team member on the command line, for what
> it's
> >> worth.
> >>
> >> > On Jan 17, 2015, at 2:39 AM, Sean Owen <so...@cloudera.com> wrote:
> >> >
> >> > Hm, this test hangs for me in IntelliJ. It could be a real problem,
> >> > and a combination of a) just recently actually enabling Java tests, b)
> >> > recent updates to the complicated Guava shading situation.
> >> >
> >> > The manifestation of the error usually suggests that something totally
> >> > failed to start (because of, say, class incompatibility errors, etc.)
> >> > Thus things hang and time out waiting for the dead component. It's
> >> > sometimes hard to get answers from the embedded component that dies
> >> > though.
> >> >
> >> > That said, it seems to pass on the command line. For example my recent
> >> > Jenkins job shows it passes:
> >> >
> >> >
> https://amplab.cs.berkeley.edu/jenkins/job/SparkPullRequestBuilder/25682/consoleFull
> >> >
> >> > I'll try to uncover more later this weekend. Thoughts welcome though.
> >> >
> >> > On Fri, Jan 16, 2015 at 8:26 PM, Andrew Musselman
> >> > <andrew.mussel...@gmail.com> wrote:
> >> >> Thanks Ted, got farther along but now have a failing test; is this a
> >> >> known
> >> >> issue?
> >> >>
> >> >> -------------------------------------------------------
> >> >> T E S T S
> >> >> -------------------------------------------------------
> >> >> Running org.apache.spark.JavaAPISuite
> >> >> Tests run: 72, Failures: 0, Errors: 1, Skipped: 0, Time elapsed:
> >> >> 123.462 sec
> >> >> <<< FAILURE! - in org.apache.spark.JavaAPISuite
> >> >> testGuavaOptional(org.apache.spark.JavaAPISuite)  Time elapsed: 106.5
> >> >> sec
> >> >> <<< ERROR!
> >> >> org.apache.spark.SparkException: Job aborted due to stage failure:
> >> >> Master
> >> >> removed our application: FAILED
> >> >>    at
> >> >>
> >> >> org.apache.spark.scheduler.DAGScheduler.org
> $apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1199)
> >> >>    at
> >> >>
> >> >>
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1188)
> >> >>    at
> >> >>
> >> >>
> org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1187)
> >> >>    at
> >> >>
> >> >>
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
> >> >>    at
> >> >> scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
> >> >>    at
> >> >>
> >> >>
> org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1187)
> >> >>    at
> >> >>
> >> >>
> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
> >> >>    at
> >> >>
> >> >>
> org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
> >> >>    at scala.Option.foreach(Option.scala:236)
> >> >>    at
> >> >>
> >> >>
> org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:697)
> >> >>    at
> >> >>
> >> >>
> org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2.applyOrElse(DAGScheduler.scala:1399)
> >> >>    at akka.actor.Actor$class.aroundReceive(Actor.scala:465)
> >> >>    at
> >> >>
> >> >>
> org.apache.spark.scheduler.DAGSchedulerEventProcessActor.aroundReceive(DAGScheduler.scala:1360)
> >> >>    at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516)
> >> >>    at akka.actor.ActorCell.invoke(ActorCell.scala:487)
> >> >>    at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:238)
> >> >>    at akka.dispatch.Mailbox.run(Mailbox.scala:220)
> >> >>    at
> >> >>
> >> >>
> akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:393)
> >> >>    at
> >> >> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
> >> >>    at
> >> >>
> >> >>
> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
> >> >>    at
> >> >>
> >> >>
> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
> >> >>    at
> >> >>
> >> >>
> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> >> >>
> >> >> Running org.apache.spark.JavaJdbcRDDSuite
> >> >> Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.846
> >> >> sec -
> >> >> in org.apache.spark.JavaJdbcRDDSuite
> >> >>
> >> >> Results :
> >> >>
> >> >>
> >> >> Tests in error:
> >> >>  JavaAPISuite.testGuavaOptional ยป Spark Job aborted due to stage
> >> >> failure:
> >> >> Maste...
> >> >>
> >> >>> On Fri, Jan 16, 2015 at 12:06 PM, Ted Yu <yuzhih...@gmail.com>
> wrote:
> >> >>>
> >> >>> Can you try doing this before running mvn ?
> >> >>>
> >> >>> export MAVEN_OPTS="-Xmx2g -XX:MaxPermSize=512M
> >> >>> -XX:ReservedCodeCacheSize=512m"
> >> >>>
> >> >>> What OS are you using ?
> >> >>>
> >> >>> Cheers
> >> >>>
> >> >>> On Fri, Jan 16, 2015 at 12:03 PM, Andrew Musselman
> >> >>> <andrew.mussel...@gmail.com> wrote:
> >> >>>>
> >> >>>> Just got the latest from Github and tried running `mvn test`; is
> this
> >> >>>> error common and do you have any advice on fixing it?
> >> >>>>
> >> >>>> Thanks!
> >> >>>>
> >> >>>> [INFO] --- scala-maven-plugin:3.2.0:compile (scala-compile-first) @
> >> >>>> spark-core_2.10 ---
> >> >>>> [WARNING] Zinc server is not available at port 3030 - reverting to
> >> >>>> normal
> >> >>>> incremental compile
> >> >>>> [INFO] Using incremental compilation
> >> >>>> [INFO] compiler plugin:
> >> >>>> BasicArtifact(org.scalamacros,paradise_2.10.4,2.0.1,null)
> >> >>>> [INFO] Compiling 400 Scala sources and 34 Java sources to
> >> >>>> /home/akm/spark/core/target/scala-2.10/classes...
> >> >>>> [WARNING]
> >> >>>>
> >> >>>>
> /home/akm/spark/core/src/main/scala/org/apache/spark/executor/TaskMetrics.scala:22:
> >> >>>> imported `DataReadMethod' is permanently hidden by definition of
> >> >>>> object
> >> >>>> DataReadMethod in package executor
> >> >>>> [WARNING] import org.apache.spark.executor.DataReadMethod
> >> >>>> [WARNING]                                  ^
> >> >>>> [WARNING]
> >> >>>>
> >> >>>>
> /home/akm/spark/core/src/main/scala/org/apache/spark/TaskState.scala:41:
> >> >>>> match may not be exhaustive.
> >> >>>> It would fail on the following input: TASK_ERROR
> >> >>>> [WARNING]   def fromMesos(mesosState: MesosTaskState): TaskState =
> >> >>>> mesosState match {
> >> >>>> [WARNING]
> ^
> >> >>>> [WARNING]
> >> >>>>
> >> >>>>
> /home/akm/spark/core/src/main/scala/org/apache/spark/scheduler/EventLoggingListener.scala:89:
> >> >>>> method isDirectory in class FileSystem is deprecated: see
> >> >>>> corresponding
> >> >>>> Javadoc for more information.
> >> >>>> [WARNING]     if (!fileSystem.isDirectory(new Path(logBaseDir))) {
> >> >>>> [WARNING]                     ^
> >> >>>> [ERROR] PermGen space -> [Help 1]
> >> >>>> [ERROR]
> >> >>>> [ERROR] To see the full stack trace of the errors, re-run Maven
> with
> >> >>>> the
> >> >>>> -e switch.
> >> >>>> [ERROR] Re-run Maven using the -X switch to enable full debug
> >> >>>> logging.
> >> >>>> [ERROR]
> >> >>>> [ERROR] For more information about the errors and possible
> solutions,
> >> >>>> please read the following articles:
> >> >>>> [ERROR] [Help 1]
> >> >>>> http://cwiki.apache.org/confluence/display/MAVEN/OutOfMemoryError
> >> >>
> >
> >
>

Reply via email to