I often see this when I first build the whole Spark project with SBT, then
modify some code and tries to build and debug within IDEA, or vice versa. A
clean rebuild can always solve this.

On Mon, Nov 3, 2014 at 11:28 AM, Patrick Wendell <pwend...@gmail.com> wrote:

> Does this happen if you clean and recompile? I've seen failures on and
> off, but haven't been able to find one that I could reproduce from a
> clean build such that we could hand it to the scala team.
>
> - Patrick
>
> On Sun, Nov 2, 2014 at 7:25 PM, Imran Rashid <im...@therashids.com> wrote:
> > I'm finding the scala compiler crashes when I compile the spark-sql
> project
> > in sbt.  This happens in both the 1.1 branch and master (full error
> > below).  The other projects build fine in sbt, and everything builds fine
> > in maven.  is there some sbt option I'm forgetting?  Any one else
> > experiencing this?
> >
> > Also, are there up-to-date instructions on how to do common dev tasks in
> > both sbt & maven?  I have only found these instructions on building with
> > maven:
> >
> > http://spark.apache.org/docs/latest/building-with-maven.html
> >
> > and some general info here:
> >
> > https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark
> >
> > but I think this doesn't walk through a lot of the steps of a typical dev
> > cycle, eg, continuous compilation, running one test, running one main
> > class, etc.  (especially since it seems like people still favor sbt for
> > dev.)  If it doesn't already exist somewhere, I could try to put
> together a
> > brief doc for how to do the basics.  (I'm returning to spark dev after a
> > little hiatus myself, and I'm hitting some stumbling blocks that are
> > probably common knowledge to everyone still dealing with it all the
> time.)
> >
> > thanks,
> > Imran
> >
> > ------------------------------
> > full crash info from sbt:
> >
> >> project sql
> > [info] Set current project to spark-sql (in build
> > file:/Users/imran/spark/spark/)
> >> compile
> > [info] Compiling 62 Scala sources to
> > /Users/imran/spark/spark/sql/catalyst/target/scala-2.10/classes...
> > [info] Compiling 45 Scala sources and 39 Java sources to
> > /Users/imran/spark/spark/sql/core/target/scala-2.10/classes...
> > [error]
> > [error]      while compiling:
> >
> /Users/imran/spark/spark/sql/core/src/main/scala/org/apache/spark/sql/types/util/DataTypeConversions.scala
> > [error]         during phase: jvm
> > [error]      library version: version 2.10.4
> > [error]     compiler version: version 2.10.4
> > [error]   reconstructed args: -classpath
> >
> /Users/imran/spark/spark/sql/core/target/scala-2.10/classes:/Users/imran/spark/spark/core/target/scala-2.10/classes:/Users/imran/spark/spark/sql/catalyst/target/scala-2.10/classes:/Users/imran/spark/spark/lib_managed/jars/hadoop-client-1.0.4.jar:/Users/imran/spark/spark/lib_managed/jars/hadoop-core-1.0.4.jar:/Users/imran/spark/spark/lib_managed/jars/xmlenc-0.52.jar:/Users/imran/spark/spark/lib_managed/jars/commons-math-2.1.jar:/Users/imran/spark/spark/lib_managed/jars/commons-configuration-1.6.jar:/Users/imran/spark/spark/lib_managed/jars/commons-collections-3.2.1.jar:/Users/imran/spark/spark/lib_managed/jars/commons-lang-2.4.jar:/Users/imran/spark/spark/lib_managed/jars/commons-logging-1.1.1.jar:/Users/imran/spark/spark/lib_managed/jars/commons-digester-1.8.jar:/Users/imran/spark/spark/lib_managed/jars/commons-beanutils-1.7.0.jar:/Users/imran/spark/spark/lib_managed/jars/commons-beanutils-core-1.8.0.jar:/Users/imran/spark/spark/lib_managed/jars/commons-net-2.2.jar:/Users/imran/spark/spark/lib_managed/jars/commons-el-1.0.jar:/Users/imran/spark/spark/lib_managed/jars/hsqldb-1.8.0.10.jar:/Users/imran/spark/spark/lib_managed/jars/oro-2.0.8.jar:/Users/imran/spark/spark/lib_managed/jars/jets3t-0.7.1.jar:/Users/imran/spark/spark/lib_managed/jars/commons-httpclient-3.1.jar:/Users/imran/spark/spark/lib_managed/bundles/curator-recipes-2.4.0.jar:/Users/imran/spark/spark/lib_managed/bundles/curator-framework-2.4.0.jar:/Users/imran/spark/spark/lib_managed/bundles/curator-client-2.4.0.jar:/Users/imran/spark/spark/lib_managed/jars/zookeeper-3.4.5.jar:/Users/imran/spark/spark/lib_managed/jars/slf4j-log4j12-1.7.5.jar:/Users/imran/spark/spark/lib_managed/bundles/log4j-1.2.17.jar:/Users/imran/spark/spark/lib_managed/jars/jline-0.9.94.jar:/Users/imran/spark/spark/lib_managed/bundles/guava-14.0.1.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-plus-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/orbits/javax.transaction-1.1.1.v201105210645.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-webapp-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-xml-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-util-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-servlet-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-security-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-server-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/orbits/javax.servlet-3.0.0.v201112011016.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-continuation-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-http-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-io-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/jars/jetty-jndi-8.1.14.v20131031.jar:/Users/imran/spark/spark/lib_managed/orbits/javax.mail.glassfish-1.4.1.v201005082020.jar:/Users/imran/spark/spark/lib_managed/orbits/javax.activation-1.1.0.v201105071233.jar:/Users/imran/spark/spark/lib_managed/jars/commons-lang3-3.3.2.jar:/Users/imran/spark/spark/lib_managed/jars/jsr305-1.3.9.jar:/Users/imran/spark/spark/lib_managed/jars/slf4j-api-1.7.5.jar:/Users/imran/spark/spark/lib_managed/jars/jul-to-slf4j-1.7.5.jar:/Users/imran/spark/spark/lib_managed/jars/jcl-over-slf4j-1.7.5.jar:/Users/imran/spark/spark/lib_managed/bundles/compress-lzf-1.0.0.jar:/Users/imran/spark/spark/lib_managed/bundles/snappy-java-1.0.5.3.jar:/Users/imran/spark/spark/lib_managed/jars/lz4-1.2.0.jar:/Users/imran/spark/spark/lib_managed/jars/chill_2.10-0.3.6.jar:/Users/imran/spark/spark/lib_managed/jars/chill-java-0.3.6.jar:/Users/imran/spark/spark/lib_managed/bundles/kryo-2.21.jar:/Users/imran/spark/spark/lib_managed/jars/reflectasm-1.07-shaded.jar:/Users/imran/spark/spark/lib_managed/jars/minlog-1.2.jar:/Users/imran/spark/spark/lib_managed/jars/objenesis-1.2.jar:/Users/imran/spark/spark/lib_managed/bundles/akka-remote_2.10-2.2.3-shaded-protobuf.jar:/Users/imran/spark/spark/lib_managed/jars/akka-actor_2.10-2.2.3-shaded-protobuf.jar:/Users/imran/spark/spark/lib_managed/bundles/config-1.0.2.jar:/Users/imran/spark/spark/lib_managed/bundles/netty-3.6.6.Final.jar:/Users/imran/spark/spark/lib_managed/jars/protobuf-java-2.4.1-shaded.jar:/Users/imran/spark/spark/lib_managed/jars/uncommons-maths-1.2.2a.jar:/Users/imran/spark/spark/lib_managed/bundles/akka-slf4j_2.10-2.2.3-shaded-protobuf.jar:/Users/imran/spark/spark/lib_managed/jars/json4s-jackson_2.10-3.2.10.jar:/Users/imran/spark/spark/lib_managed/jars/json4s-core_2.10-3.2.10.jar:/Users/imran/spark/spark/lib_managed/jars/json4s-ast_2.10-3.2.10.jar:/Users/imran/spark/spark/lib_managed/jars/paranamer-2.6.jar:/Users/imran/spark/spark/lib_managed/jars/scalap-2.10.0.jar:/Users/imran/spark/spark/lib_managed/bundles/jackson-databind-2.3.1.jar:/Users/imran/spark/spark/lib_managed/bundles/jackson-annotations-2.3.0.jar:/Users/imran/spark/spark/lib_managed/bundles/jackson-core-2.3.1.jar:/Users/imran/spark/spark/lib_managed/jars/colt-1.2.0.jar:/Users/imran/spark/spark/lib_managed/jars/concurrent-1.3.4.jar:/Users/imran/spark/spark/lib_managed/jars/mesos-0.18.1-shaded-protobuf.jar:/Users/imran/spark/spark/lib_managed/jars/netty-all-4.0.23.Final.jar:/Users/imran/spark/spark/lib_managed/jars/stream-2.7.0.jar:/Users/imran/spark/spark/lib_managed/bundles/metrics-core-3.0.0.jar:/Users/imran/spark/spark/lib_managed/bundles/metrics-jvm-3.0.0.jar:/Users/imran/spark/spark/lib_managed/bundles/metrics-json-3.0.0.jar:/Users/imran/spark/spark/lib_managed/bundles/metrics-graphite-3.0.0.jar:/Users/imran/spark/spark/lib_managed/jars/tachyon-client-0.5.0.jar:/Users/imran/spark/spark/lib_managed/jars/tachyon-0.5.0.jar:/Users/imran/spark/spark/lib_managed/jars/commons-io-2.4.jar:/Users/imran/spark/spark/lib_managed/jars/pyrolite-2.0.1.jar:/Users/imran/spark/spark/lib_managed/jars/py4j-0.8.2.1.jar:/Users/imran/.sbt/boot/scala-2.10.4/lib/scala-compiler.jar:/Users/imran/.sbt/boot/scala-2.10.4/lib/scala-reflect.jar:/Users/imran/spark/spark/lib_managed/jars/quasiquotes_2.10-2.0.1.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-column-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-common-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-encoding-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-generator-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/commons-codec-1.5.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-hadoop-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-format-2.0.0.jar:/Users/imran/spark/spark/lib_managed/jars/parquet-jackson-1.4.3.jar:/Users/imran/spark/spark/lib_managed/jars/jackson-mapper-asl-1.9.11
> >
> .jar:/Users/imran/spark/spark/lib_managed/jars/jackson-core-asl-1.9.11.jar
> > -deprecation -feature
> > -P:genjavadoc:out=/Users/imran/spark/spark/sql/core/target/java
> >
> -Xplugin:/Users/imran/spark/spark/lib_managed/jars/genjavadoc-plugin_2.10.4-0.7.jar
> > -bootclasspath
> >
> /Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/resources.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/rt.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/sunrsasign.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/jsse.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/jce.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/charsets.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/lib/jfr.jar:/Library/Java/JavaVirtualMachines/jdk1.7.0_51.jdk/Contents/Home/jre/classes:/Users/imran/.sbt/boot/scala-2.10.4/lib/scala-library.jar
> > -unchecked -language:postfixOps
> > [error]
> > [error]   last tree to typer:
> > Literal(Constant(org.apache.spark.sql.catalyst.types.PrimitiveType))
> > [error]               symbol: null
> > [error]    symbol definition: null
> > [error]                  tpe:
> > Class(classOf[org.apache.spark.sql.catalyst.types.PrimitiveType])
> > [error]        symbol owners:
> > [error]       context owners: anonymous class anonfun$asScalaDataType$1
> ->
> > package util
> > [error]
> > [error] == Enclosing template or block ==
> > [error]
> > [error] Template( // val <local $anonfun>: <notype>,
> > tree.tpe=org.apache.spark.sql.types.util.anonfun$asScalaDataType$1
> > [error]   "scala.runtime.AbstractFunction1", "scala.Serializable" //
> parents
> > [error]   ValDef(
> > [error]     private
> > [error]     "_"
> > [error]     <tpt>
> > [error]     <empty>
> > [error]   )
> > [error]   // 3 statements
> > [error]   DefDef( // final def apply(javaStructField:
> > org.apache.spark.sql.api.java.StructField):
> > org.apache.spark.sql.catalyst.types.StructField
> > [error]     <method> final <triedcooking>
> > [error]     "apply"
> > [error]     []
> > [error]     // 1 parameter list
> > [error]     ValDef( // javaStructField:
> > org.apache.spark.sql.api.java.StructField
> > [error]       <param> <synthetic> <triedcooking>
> > [error]       "javaStructField"
> > [error]       <tpt> // tree.tpe=org.apache.spark.sql.api.java.StructField
> > [error]       <empty>
> > [error]     )
> > [error]     <tpt> //
> > tree.tpe=org.apache.spark.sql.catalyst.types.StructField
> > [error]     Apply( // def asScalaStructField(javaStructField:
> > org.apache.spark.sql.api.java.StructField):
> > org.apache.spark.sql.catalyst.types.StructField in object
> > DataTypeConversions,
> > tree.tpe=org.apache.spark.sql.catalyst.types.StructField
> > [error]       DataTypeConversions.this."asScalaStructField" // def
> > asScalaStructField(javaStructField:
> > org.apache.spark.sql.api.java.StructField):
> > org.apache.spark.sql.catalyst.types.StructField in object
> > DataTypeConversions, tree.tpe=(javaStructField:
> >
> org.apache.spark.sql.api.java.StructField)org.apache.spark.sql.catalyst.types.StructField
> > [error]       "javaStructField" // javaStructField:
> > org.apache.spark.sql.api.java.StructField,
> > tree.tpe=org.apache.spark.sql.api.java.StructField
> > [error]     )
> > [error]   )
> > [error]   DefDef( // final def apply(v1: Object): Object
> > [error]     <method> final <bridge>
> > [error]     "apply"
> > [error]     []
> > [error]     // 1 parameter list
> > [error]     ValDef( // v1: Object
> > [error]       <param> <triedcooking>
> > [error]       "v1"
> > [error]       <tpt> // tree.tpe=Object
> > [error]       <empty>
> > [error]     )
> > [error]     <tpt> // tree.tpe=Object
> > [error]     Apply( // final def apply(javaStructField:
> > org.apache.spark.sql.api.java.StructField):
> > org.apache.spark.sql.catalyst.types.StructField,
> > tree.tpe=org.apache.spark.sql.catalyst.types.StructField
> > [error]       DataTypeConversions$$anonfun$asScalaDataType$1.this."apply"
> > // final def apply(javaStructField:
> > org.apache.spark.sql.api.java.StructField):
> > org.apache.spark.sql.catalyst.types.StructField,
> tree.tpe=(javaStructField:
> >
> org.apache.spark.sql.api.java.StructField)org.apache.spark.sql.catalyst.types.StructField
> > [error]       Apply( // final def $asInstanceOf[T0 >: ? <: ?](): T0 in
> > class Object, tree.tpe=org.apache.spark.sql.api.java.StructField
> > [error]         TypeApply( // final def $asInstanceOf[T0 >: ? <: ?](): T0
> > in class Object, tree.tpe=()org.apache.spark.sql.api.java.StructField
> > [error]           "v1"."$asInstanceOf" // final def $asInstanceOf[T0 >: ?
> > <: ?](): T0 in class Object, tree.tpe=[T0 >: ? <: ?]()T0
> > [error]           <tpt> //
> > tree.tpe=org.apache.spark.sql.api.java.StructField
> > [error]         )
> > [error]         Nil
> > [error]       )
> > [error]     )
> > [error]   )
> > [error]   DefDef( // def <init>():
> > org.apache.spark.sql.types.util.anonfun$asScalaDataType$1
> > [error]     <method> <triedcooking>
> > [error]     "<init>"
> > [error]     []
> > [error]     List(Nil)
> > [error]     <tpt> //
> > tree.tpe=org.apache.spark.sql.types.util.anonfun$asScalaDataType$1
> > [error]     Block( // tree.tpe=Unit
> > [error]       Apply( // def <init>(): scala.runtime.AbstractFunction1 in
> > class AbstractFunction1, tree.tpe=scala.runtime.AbstractFunction1
> > [error]
> > DataTypeConversions$$anonfun$asScalaDataType$1.super."<init>" // def
> > <init>(): scala.runtime.AbstractFunction1 in class AbstractFunction1,
> > tree.tpe=()scala.runtime.AbstractFunction1
> > [error]         Nil
> > [error]       )
> > [error]       ()
> > [error]     )
> > [error]   )
> > [error] )
> > [error]
> > [error] == Expanded type of tree ==
> > [error]
> > [error] ConstantType(
> > [error]   value =
> > Constant(org.apache.spark.sql.catalyst.types.PrimitiveType)
> > [error] )
> > [error]
> > [error] uncaught exception during compilation: java.lang.AssertionError
> > [trace] Stack trace suppressed: run last sql/compile:compile for the full
> > output.
> > [error] (sql/compile:compile) java.lang.AssertionError: assertion failed:
> > List(object package$DebugNode, object package$DebugNode)
> > [error] Total time: 23 s, completed Nov 2, 2014 1:00:37 PM
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
> For additional commands, e-mail: dev-h...@spark.apache.org
>
>

Reply via email to