Is the Scala version in Intellij the same as the one used by sbt ?

Cheers

On Tue, Nov 17, 2015 at 6:45 PM, 金国栋 <scu...@gmail.com> wrote:

> Hi!
>
> I tried to build spark source code from github, and I successfully built
> it from command line using `*sbt/sbt assembly*`. While I encountered an
> error when compiling the project in Intellij IDEA(V14.1.5).
>
>
> The error log is below:
> *Error:scala: *
> *     while compiling:
> /Users/ray/Documents/P01_Project/Spark-Github/spark/sql/core/src/main/scala/org/apache/spark/sql/util/QueryExecutionListener.scala*
> *        during phase: jvm*
>      library version: version 2.10.5
>     compiler version: version 2.10.5
>   reconstructed args: -nobootcp -javabootclasspath : -deprecation -feature
> -classpath
> /Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/lib/ant-javafx.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/lib/dt.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/lib/javafx-mx.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/lib/jconsole.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/lib/sa-jdi.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/lib/tools.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/charsets.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/deploy.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/htmlconverter.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/javaws.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/jce.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/jfr.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/jfxswt.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/jsse.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/management-agent.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/plugin.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/resources.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/rt.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/cldrdata.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/dnsns.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/jfxrt.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/localedata.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/nashorn.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/sunec.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/sunjce_provider.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/sunpkcs11.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/zipfs.jar:/Users/ray/Documents/P01_Project/Spark-Github/spark/sql/core/target/scala-2.10/classes:/Users/ray/Documents/P01_Project/Spark-Github/spark/core/target/scala-2.10/classes:/Users/ray/.m2/repository/org/apache/avro/avro-mapred/1.7.7/avro-mapred-1.7.7-hadoop2.jar:/Users/ray/.m2/repository/org/apache/avro/avro-ipc/1.7.7/avro-ipc-1.7.7.jar:/Users/ray/.m2/repository/org/apache/avro/avro-ipc/1.7.7/avro-ipc-1.7.7-tests.jar:/Users/ray/.m2/repository/com/twitter/chill_2.10/0.5.0/chill_2.10-0.5.0.jar:/Users/ray/.m2/repository/com/esotericsoftware/kryo/kryo/2.21/kryo-2.21.jar:/Users/ray/.m2/repository/com/esotericsoftware/reflectasm/reflectasm/1.07/reflectasm-1.07-shaded.jar:/Users/ray/.m2/repository/com/esotericsoftware/minlog/minlog/1.2/minlog-1.2.jar:/Users/ray/.m2/repository/com/twitter/chill-java/0.5.0/chill-java-0.5.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-client/2.2.0/hadoop-client-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-common/2.2.0/hadoop-common-2.2.0.jar:/Users/ray/.m2/repository/commons-cli/commons-cli/1.2/commons-cli-1.2.jar:/Users/ray/.m2/repository/org/apache/commons/commons-math/2.1/commons-math-2.1.jar:/Users/ray/.m2/repository/xmlenc/xmlenc/0.52/xmlenc-0.52.jar:/Users/ray/.m2/repository/commons-configuration/commons-configuration/1.6/commons-configuration-1.6.jar:/Users/ray/.m2/repository/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar:/Users/ray/.m2/repository/commons-digester/commons-digester/1.8/commons-digester-1.8.jar:/Users/ray/.m2/repository/commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar:/Users/ray/.m2/repository/commons-beanutils/commons-beanutils-core/1.8.0/commons-beanutils-core-1.8.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-auth/2.2.0/hadoop-auth-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-hdfs/2.2.0/hadoop-hdfs-2.2.0.jar:/Users/ray/.m2/repository/org/mortbay/jetty/jetty-util/6.1.26/jetty-util-6.1.26.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-mapreduce-client-app/2.2.0/hadoop-mapreduce-client-app-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-mapreduce-client-common/2.2.0/hadoop-mapreduce-client-common-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-yarn-client/2.2.0/hadoop-yarn-client-2.2.0.jar:/Users/ray/.m2/repository/com/google/inject/guice/3.0/guice-3.0.jar:/Users/ray/.m2/repository/javax/inject/javax.inject/1/javax.inject-1.jar:/Users/ray/.m2/repository/aopalliance/aopalliance/1.0/aopalliance-1.0.jar:/Users/ray/.m2/repository/com/sun/jersey/jersey-test-framework/jersey-test-framework-grizzly2/1.9/jersey-test-framework-grizzly2-1.9.jar:/Users/ray/.m2/repository/com/sun/jersey/jersey-test-framework/jersey-test-framework-core/1.9/jersey-test-framework-core-1.9.jar:/Users/ray/.m2/repository/javax/servlet/javax.servlet-api/3.0.1/javax.servlet-api-3.0.1.jar:/Users/ray/.m2/repository/com/sun/jersey/jersey-client/1.9/jersey-client-1.9.jar:/Users/ray/.m2/repository/com/sun/jersey/jersey-grizzly2/1.9/jersey-grizzly2-1.9.jar:/Users/ray/.m2/repository/org/glassfish/grizzly/grizzly-http/2.1.2/grizzly-http-2.1.2.jar:/Users/ray/.m2/repository/org/glassfish/grizzly/grizzly-framework/2.1.2/grizzly-framework-2.1.2.jar:/Users/ray/.m2/repository/org/glassfish/gmbal/gmbal-api-only/3.0.0-b023/gmbal-api-only-3.0.0-b023.jar:/Users/ray/.m2/repository/org/glassfish/external/management-api/3.0.0-b012/management-api-3.0.0-b012.jar:/Users/ray/.m2/repository/org/glassfish/grizzly/grizzly-http-server/2.1.2/grizzly-http-server-2.1.2.jar:/Users/ray/.m2/repository/org/glassfish/grizzly/grizzly-rcm/2.1.2/grizzly-rcm-2.1.2.jar:/Users/ray/.m2/repository/org/glassfish/grizzly/grizzly-http-servlet/2.1.2/grizzly-http-servlet-2.1.2.jar:/Users/ray/.m2/repository/org/glassfish/javax.servlet/3.1/javax.servlet-3.1.jar:/Users/ray/.m2/repository/com/sun/jersey/jersey-json/1.9/jersey-json-1.9.jar:/Users/ray/.m2/repository/org/codehaus/jettison/jettison/1.1/jettison-1.1.jar:/Users/ray/.m2/repository/com/sun/xml/bind/jaxb-impl/2.2.3-1/jaxb-impl-2.2.3-1.jar:/Users/ray/.m2/repository/javax/xml/bind/jaxb-api/2.2.2/jaxb-api-2.2.2.jar:/Users/ray/.m2/repository/javax/activation/activation/1.1/activation-1.1.jar:/Users/ray/.m2/repository/org/codehaus/jackson/jackson-jaxrs/1.9.13/jackson-jaxrs-1.9.13.jar:/Users/ray/.m2/repository/org/codehaus/jackson/jackson-xc/1.9.13/jackson-xc-1.9.13.jar:/Users/ray/.m2/repository/com/sun/jersey/contribs/jersey-guice/1.9/jersey-guice-1.9.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-yarn-server-common/2.2.0/hadoop-yarn-server-common-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-mapreduce-client-shuffle/2.2.0/hadoop-mapreduce-client-shuffle-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-yarn-api/2.2.0/hadoop-yarn-api-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-mapreduce-client-core/2.2.0/hadoop-mapreduce-client-core-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-yarn-common/2.2.0/hadoop-yarn-common-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-mapreduce-client-jobclient/2.2.0/hadoop-mapreduce-client-jobclient-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-annotations/2.2.0/hadoop-annotations-2.2.0.jar:/Users/ray/Documents/P01_Project/Spark-Github/spark/launcher/target/scala-2.10/classes:/Users/ray/Documents/P01_Project/Spark-Github/spark/network/common/target/scala-2.10/classes:/Users/ray/.m2/repository/com/google/guava/guava/14.0.1/guava-14.0.1.jar:/Users/ray/Documents/P01_Project/Spark-Github/spark/network/shuffle/target/scala-2.10/classes:/Users/ray/.m2/repository/org/fusesource/leveldbjni/leveldbjni-all/1.8/leveldbjni-all-1.8.jar:/Users/ray/Documents/P01_Project/Spark-Github/spark/unsafe/target/scala-2.10/classes:/Users/ray/.m2/repository/net/java/dev/jets3t/jets3t/0.7.1/jets3t-0.7.1.jar:/Users/ray/.m2/repository/commons-httpclient/commons-httpclient/3.1/commons-httpclient-3.1.jar:/Users/ray/.m2/repository/org/apache/curator/curator-recipes/2.4.0/curator-recipes-2.4.0.jar:/Users/ray/.m2/repository/org/apache/curator/curator-framework/2.4.0/curator-framework-2.4.0.jar:/Users/ray/.m2/repository/org/apache/curator/curator-client/2.4.0/curator-client-2.4.0.jar:/Users/ray/.m2/repository/org/apache/zookeeper/zookeeper/3.4.5/zookeeper-3.4.5.jar:/Users/ray/.m2/repository/jline/jline/0.9.94/jline-0.9.94.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-plus/8.1.14.v20131031/jetty-plus-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/orbit/javax.transaction/1.1.1.v201105210645/javax.transaction-1.1.1.v201105210645.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-webapp/8.1.14.v20131031/jetty-webapp-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-xml/8.1.14.v20131031/jetty-xml-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-jndi/8.1.14.v20131031/jetty-jndi-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/orbit/javax.mail.glassfish/1.4.1.v201005082020/javax.mail.glassfish-1.4.1.v201005082020.jar:/Users/ray/.m2/repository/org/eclipse/jetty/orbit/javax.activation/1.1.0.v201105071233/javax.activation-1.1.0.v201105071233.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-security/8.1.14.v20131031/jetty-security-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-util/8.1.14.v20131031/jetty-util-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-server/8.1.14.v20131031/jetty-server-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-http/8.1.14.v20131031/jetty-http-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-io/8.1.14.v20131031/jetty-io-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-continuation/8.1.14.v20131031/jetty-continuation-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-servlet/8.1.14.v20131031/jetty-servlet-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/orbit/javax.servlet/3.0.0.v201112011016/javax.servlet-3.0.0.v201112011016.jar:/Users/ray/.m2/repository/org/apache/commons/commons-lang3/3.3.2/commons-lang3-3.3.2.jar:/Users/ray/.m2/repository/org/apache/commons/commons-math3/3.4.1/commons-math3-3.4.1.jar:/Users/ray/.m2/repository/com/google/code/findbugs/jsr305/1.3.9/jsr305-1.3.9.jar:/Users/ray/.m2/repository/org/slf4j/slf4j-api/1.7.10/slf4j-api-1.7.10.jar:/Users/ray/.m2/repository/org/slf4j/jul-to-slf4j/1.7.10/jul-to-slf4j-1.7.10.jar:/Users/ray/.m2/repository/org/slf4j/jcl-over-slf4j/1.7.10/jcl-over-slf4j-1.7.10.jar:/Users/ray/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar:/Users/ray/.m2/repository/org/slf4j/slf4j-log4j12/1.7.10/slf4j-log4j12-1.7.10.jar:/Users/ray/.m2/repository/com/ning/compress-lzf/1.0.3/compress-lzf-1.0.3.jar:/Users/ray/.m2/repository/org/xerial/snappy/snappy-java/1.1.2/snappy-java-1.1.2.jar:/Users/ray/.m2/repository/net/jpountz/lz4/lz4/1.3.0/lz4-1.3.0.jar:/Users/ray/.m2/repository/org/roaringbitmap/RoaringBitmap/0.4.5/RoaringBitmap-0.4.5.jar:/Users/ray/.m2/repository/commons-net/commons-net/2.2/commons-net-2.2.jar:/Users/ray/.m2/repository/com/typesafe/akka/akka-remote_2.10/2.3.11/akka-remote_2.10-2.3.11.jar:/Users/ray/.m2/repository/com/typesafe/akka/akka-actor_2.10/2.3.11/akka-actor_2.10-2.3.11.jar:/Users/ray/.m2/repository/com/typesafe/config/1.2.1/config-1.2.1.jar:/Users/ray/.m2/repository/io/netty/netty/3.8.0.Final/netty-3.8.0.Final.jar:/Users/ray/.m2/repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar:/Users/ray/.m2/repository/org/uncommons/maths/uncommons-maths/1.2.2a/uncommons-maths-1.2.2a.jar:/Users/ray/.m2/repository/com/typesafe/akka/akka-slf4j_2.10/2.3.11/akka-slf4j_2.10-2.3.11.jar:/Users/ray/.m2/repository/org/scala-lang/scala-library/2.10.5/scala-library-2.10.5.jar:/Users/ray/.m2/repository/org/json4s/json4s-jackson_2.10/3.2.10/json4s-jackson_2.10-3.2.10.jar:/Users/ray/.m2/repository/org/json4s/json4s-core_2.10/3.2.10/json4s-core_2.10-3.2.10.jar:/Users/ray/.m2/repository/org/json4s/json4s-ast_2.10/3.2.10/json4s-ast_2.10-3.2.10.jar:/Users/ray/.m2/repository/org/scala-lang/scalap/2.10.5/scalap-2.10.5.jar:/Users/ray/.m2/repository/org/scala-lang/scala-compiler/2.10.5/scala-compiler-2.10.5.jar:/Users/ray/.m2/repository/com/sun/jersey/jersey-server/1.9/jersey-server-1.9.jar:/Users/ray/.m2/repository/asm/asm/3.1/asm-3.1.jar:/Users/ray/.m2/repository/com/sun/jersey/jersey-core/1.9/jersey-core-1.9.jar:/Users/ray/.m2/repository/org/apache/mesos/mesos/0.21.1/mesos-0.21.1-shaded-protobuf.jar:/Users/ray/.m2/repository/io/netty/netty-all/4.0.29.Final/netty-all-4.0.29.Final.jar:/Users/ray/.m2/repository/com/clearspring/analytics/stream/2.7.0/stream-2.7.0.jar:/Users/ray/.m2/repository/io/dropwizard/metrics/metrics-core/3.1.2/metrics-core-3.1.2.jar:/Users/ray/.m2/repository/io/dropwizard/metrics/metrics-jvm/3.1.2/metrics-jvm-3.1.2.jar:/Users/ray/.m2/repository/io/dropwizard/metrics/metrics-json/3.1.2/metrics-json-3.1.2.jar:/Users/ray/.m2/repository/io/dropwizard/metrics/metrics-graphite/3.1.2/metrics-graphite-3.1.2.jar:/Users/ray/.m2/repository/com/fasterxml/jackson/module/jackson-module-scala_2.10/2.4.4/jackson-module-scala_2.10-2.4.4.jar:/Users/ray/.m2/repository/com/thoughtworks/paranamer/paranamer/2.6/paranamer-2.6.jar:/Users/ray/.m2/repository/org/apache/ivy/ivy/2.4.0/ivy-2.4.0.jar:/Users/ray/.m2/repository/oro/oro/2.0.8/oro-2.0.8.jar:/Users/ray/.m2/repository/org/tachyonproject/tachyon-client/0.8.1/tachyon-client-0.8.1.jar:/Users/ray/.m2/repository/commons-lang/commons-lang/2.6/commons-lang-2.6.jar:/Users/ray/.m2/repository/commons-io/commons-io/2.4/commons-io-2.4.jar:/Users/ray/.m2/repository/org/tachyonproject/tachyon-underfs-hdfs/0.8.1/tachyon-underfs-hdfs-0.8.1.jar:/Users/ray/.m2/repository/org/tachyonproject/tachyon-underfs-s3/0.8.1/tachyon-underfs-s3-0.8.1.jar:/Users/ray/.m2/repository/org/tachyonproject/tachyon-underfs-local/0.8.1/tachyon-underfs-local-0.8.1.jar:/Users/ray/.m2/repository/net/razorvine/pyrolite/4.9/pyrolite-4.9.jar:/Users/ray/.m2/repository/net/sf/py4j/py4j/0.9/py4j-0.9.jar:/Users/ray/Documents/P01_Project/Spark-Github/spark/sql/catalyst/target/scala-2.10/classes:/Users/ray/.m2/repository/org/scala-lang/scala-reflect/2.10.5/scala-reflect-2.10.5.jar:/Users/ray/.m2/repository/org/codehaus/janino/janino/2.7.8/janino-2.7.8.jar:/Users/ray/.m2/repository/org/codehaus/janino/commons-compiler/2.7.8/commons-compiler-2.7.8.jar:/Users/ray/.m2/repository/org/apache/parquet/parquet-column/1.7.0/parquet-column-1.7.0.jar:/Users/ray/.m2/repository/org/apache/parquet/parquet-common/1.7.0/parquet-common-1.7.0.jar:/Users/ray/.m2/repository/org/apache/parquet/parquet-encoding/1.7.0/parquet-encoding-1.7.0.jar:/Users/ray/.m2/repository/org/apache/parquet/parquet-generator/1.7.0/parquet-generator-1.7.0.jar:/Users/ray/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar:/Users/ray/.m2/repository/org/apache/parquet/parquet-hadoop/1.7.0/parquet-hadoop-1.7.0.jar:/Users/ray/.m2/repository/org/apache/parquet/parquet-format/2.3.0-incubating/parquet-format-2.3.0-incubating.jar:/Users/ray/.m2/repository/org/apache/parquet/parquet-jackson/1.7.0/parquet-jackson-1.7.0.jar:/Users/ray/.m2/repository/org/codehaus/jackson/jackson-mapper-asl/1.9.13/jackson-mapper-asl-1.9.13.jar:/Users/ray/.m2/repository/org/codehaus/jackson/jackson-core-asl/1.9.13/jackson-core-asl-1.9.13.jar:/Users/ray/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.4.4/jackson-databind-2.4.4.jar:/Users/ray/.m2/repository/com/fasterxml/jackson/core/jackson-annotations/2.4.4/jackson-annotations-2.4.4.jar:/Users/ray/.m2/repository/com/fasterxml/jackson/core/jackson-core/2.4.4/jackson-core-2.4.4.jar:/Users/ray/.m2/repository/org/apache/avro/avro/1.7.7/avro-1.7.7.jar:/Users/ray/.m2/repository/org/apache/commons/commons-compress/1.4.1/commons-compress-1.4.1.jar:/Users/ray/.m2/repository/org/tukaani/xz/1.0/xz-1.0.jar:/Users/ray/.m2/repository/org/objenesis/objenesis/1.0/objenesis-1.0.jar:/Users/ray/.m2/repository/org/spark-project/spark/unused/1.0.0/unused-1.0.0.jar
> -unchecked
>   last tree to typer:
> Literal(Constant(org.apache.spark.sql.test.ExamplePoint))
>               symbol: null
>    symbol definition: null
>                  tpe:
> Class(classOf[org.apache.spark.sql.test.ExamplePoint])
>        symbol owners:
>       context owners: class ExamplePointUDT -> package test
> == Enclosing template or block ==
> Template( // val <local ExamplePointUDT>: <notype> in class
> ExamplePointUDT, tree.tpe=org.apache.spark.sql.test.ExamplePointUDT
>   "org.apache.spark.sql.types.UserDefinedType" // parents
>   ValDef(
>     private
>     "_"
>     <tpt>
>     <empty>
>   )
>   // 11 statements
>   DefDef( // override def sqlType(): org.apache.spark.sql.types.DataType
> in class ExamplePointUDT
>     <method> override
>     "sqlType"
>     []
>     List(Nil)
>     <tpt> // tree.tpe=org.apache.spark.sql.types.DataType
>     Apply( // def <init>(elementType:
> org.apache.spark.sql.types.DataType,containsNull: Boolean):
> org.apache.spark.sql.types.ArrayType in class ArrayType,
> tree.tpe=org.apache.spark.sql.types.ArrayType
>       new org.apache.spark.sql.types.ArrayType."<init>" // def
> <init>(elementType: org.apache.spark.sql.types.DataType,containsNull:
> Boolean): org.apache.spark.sql.types.ArrayType in class ArrayType,
> tree.tpe=(elementType: org.apache.spark.sql.types.DataType, containsNull:
> Boolean)org.apache.spark.sql.types.ArrayType
>       // 2 arguments
>       "org"."apache"."spark"."sql"."types"."DoubleType" // case object
> DoubleType in package types,
> tree.tpe=org.apache.spark.sql.types.DoubleType.type
>       false
>     )
>   )
>   DefDef( // override def pyUDT(): String in class ExamplePointUDT
>     <method> override
>     "pyUDT"
>     []
>     List(Nil)
>     <tpt> // tree.tpe=String
>     "pyspark.sql.tests.ExamplePointUDT"
>   )
>   DefDef( // override def serialize(obj: Object):
> org.apache.spark.sql.catalyst.util.GenericArrayData in class ExamplePointUDT
>     <method> override <triedcooking>
>     "serialize"
>     []
>     // 1 parameter list
>     ValDef( // obj: Object
>       <param> <triedcooking>
>       "obj"
>       <tpt> // tree.tpe=Object
>       <empty>
>     )
>     <tpt> // tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>     Block( // tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>       // 3 statements
>       ValDef( // case val x1: Object
>         case <synthetic> <triedcooking>
>         "x1"
>         <tpt> // tree.tpe=Object
>         "obj" // obj: Object, tree.tpe=Object
>       )
>       LabelDef( // case def case5():
> org.apache.spark.sql.catalyst.util.GenericArrayData,
> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>         ()
>         If( // tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>           Apply( // final def $isInstanceOf[T0 >: ? <: ?](): Boolean in
> class Object, tree.tpe=Boolean
>             TypeApply( // final def $isInstanceOf[T0 >: ? <: ?](): Boolean
> in class Object, tree.tpe=()Boolean
>               "x1"."$isInstanceOf" // final def $isInstanceOf[T0 >: ? <:
> ?](): Boolean in class Object, tree.tpe=[T0 >: ? <: ?]()Boolean
>               <tpt> // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>             )
>             Nil
>           )
>           Block( //
> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>             ValDef( // val x2: org.apache.spark.sql.test.ExamplePoint
>               <synthetic> <triedcooking>
>               "x2"
>               <tpt> // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>               Typed( // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>                 Apply( // final def $asInstanceOf[T0 >: ? <: ?](): T0 in
> class Object, tree.tpe=org.apache.spark.sql.test.ExamplePoint
>                   TypeApply( // final def $asInstanceOf[T0 >: ? <: ?]():
> T0 in class Object, tree.tpe=()org.apache.spark.sql.test.ExamplePoint
>                     "x1"."$asInstanceOf" // final def $asInstanceOf[T0 >:
> ? <: ?](): T0 in class Object, tree.tpe=[T0 >: ? <: ?]()T0
>                     <tpt> //
> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>                   )
>                   Nil
>                 )
>                 <tpt> // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>               )
>             )
>             Apply( // case def matchEnd4(x:
> org.apache.spark.sql.catalyst.util.GenericArrayData):
> org.apache.spark.sql.catalyst.util.GenericArrayData,
> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>               "matchEnd4" // case def matchEnd4(x:
> org.apache.spark.sql.catalyst.util.GenericArrayData):
> org.apache.spark.sql.catalyst.util.GenericArrayData, tree.tpe=(x:
> org.apache.spark.sql.catalyst.util.GenericArrayData)org.apache.spark.sql.catalyst.util.GenericArrayData
>               Block( //
> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>                 // 3 statements
>                 ValDef( // val output: Array[Object]
>                   <triedcooking>
>                   "output"
>                   <tpt> // tree.tpe=Array[Object]
>                   Apply( // def <init>(_length: Int): Array[T] in class
> Array, tree.tpe=Array[Object]
>                     new Array[Object]."<init>" // def <init>(_length:
> Int): Array[T] in class Array, tree.tpe=(_length: Int)Array[Object]
>                     2
>                   )
>                 )
>                 Apply( // def update(i: Int,x: T): Unit in class Array,
> tree.tpe=Unit
>                   "output"."update" // def update(i: Int,x: T): Unit in
> class Array, tree.tpe=(i: Int, x: Object)Unit
>                   // 2 arguments
>                   0
>                   Apply( // def box(x: Double): Double in object Double,
> tree.tpe=Object
>                     "scala"."Double"."box" // def box(x: Double): Double
> in object Double, tree.tpe=(x: Double)Double
>                     Apply( // val x(): Double in class ExamplePoint,
> tree.tpe=Double
>                       "x2"."x" // val x(): Double in class ExamplePoint,
> tree.tpe=()Double
>                       Nil
>                     )
>                   )
>                 )
>                 Apply( // def update(i: Int,x: T): Unit in class Array,
> tree.tpe=Unit
>                   "output"."update" // def update(i: Int,x: T): Unit in
> class Array, tree.tpe=(i: Int, x: Object)Unit
>                   // 2 arguments
>                   1
>                   Apply( // def box(x: Double): Double in object Double,
> tree.tpe=Object
>                     "scala"."Double"."box" // def box(x: Double): Double
> in object Double, tree.tpe=(x: Double)Double
>                     Apply( // val y(): Double in class ExamplePoint,
> tree.tpe=Double
>                       "x2"."y" // val y(): Double in class ExamplePoint,
> tree.tpe=()Double
>                       Nil
>                     )
>                   )
>                 )
>                 Apply( // def <init>(array: Array[Object]):
> org.apache.spark.sql.catalyst.util.GenericArrayData in class
> GenericArrayData,
> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>                   new
> org.apache.spark.sql.catalyst.util.GenericArrayData."<init>" // def
> <init>(array: Array[Object]):
> org.apache.spark.sql.catalyst.util.GenericArrayData in class
> GenericArrayData, tree.tpe=(array:
> Array[Object])org.apache.spark.sql.catalyst.util.GenericArrayData
>                   "output" // val output: Array[Object],
> tree.tpe=Array[Object]
>                 )
>               )
>             )
>           )
>           Apply( // case def case6():
> org.apache.spark.sql.catalyst.util.GenericArrayData,
> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>             "case6" // case def case6():
> org.apache.spark.sql.catalyst.util.GenericArrayData,
> tree.tpe=()org.apache.spark.sql.catalyst.util.GenericArrayData
>             Nil
>           )
>         )
>       )
>       LabelDef( // case def case6():
> org.apache.spark.sql.catalyst.util.GenericArrayData,
> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>         ()
>         Apply( // case def matchEnd4(x:
> org.apache.spark.sql.catalyst.util.GenericArrayData):
> org.apache.spark.sql.catalyst.util.GenericArrayData,
> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>           "matchEnd4" // case def matchEnd4(x:
> org.apache.spark.sql.catalyst.util.GenericArrayData):
> org.apache.spark.sql.catalyst.util.GenericArrayData, tree.tpe=(x:
> org.apache.spark.sql.catalyst.util.GenericArrayData)org.apache.spark.sql.catalyst.util.GenericArrayData
>           Throw( // tree.tpe=Nothing
>             Apply( // def <init>(obj: Object): MatchError in class
> MatchError, tree.tpe=MatchError
>               new MatchError."<init>" // def <init>(obj: Object):
> MatchError in class MatchError, tree.tpe=(obj: Object)MatchError
>               "x1" // case val x1: Object, tree.tpe=Object
>             )
>           )
>         )
>       )
>       LabelDef( // case def matchEnd4(x:
> org.apache.spark.sql.catalyst.util.GenericArrayData):
> org.apache.spark.sql.catalyst.util.GenericArrayData,
> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>         "x" // x: org.apache.spark.sql.catalyst.util.GenericArrayData,
> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>         "x" // x: org.apache.spark.sql.catalyst.util.GenericArrayData,
> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>       )
>     )
>   )
>   DefDef( // override def deserialize(datum: Object):
> org.apache.spark.sql.test.ExamplePoint in class ExamplePointUDT
>     <method> override <triedcooking>
>     "deserialize"
>     []
>     // 1 parameter list
>     ValDef( // datum: Object
>       <param> <triedcooking>
>       "datum"
>       <tpt> // tree.tpe=Object
>       <empty>
>     )
>     <tpt> // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>     Block( // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>       // 3 statements
>       ValDef( // case val x1: Object
>         case <synthetic> <triedcooking>
>         "x1"
>         <tpt> // tree.tpe=Object
>         "datum" // datum: Object, tree.tpe=Object
>       )
>       LabelDef( // case def case5():
> org.apache.spark.sql.test.ExamplePoint,
> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>         ()
>         If( // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>           Apply( // final def $isInstanceOf[T0 >: ? <: ?](): Boolean in
> class Object, tree.tpe=Boolean
>             TypeApply( // final def $isInstanceOf[T0 >: ? <: ?](): Boolean
> in class Object, tree.tpe=()Boolean
>               "x1"."$isInstanceOf" // final def $isInstanceOf[T0 >: ? <:
> ?](): Boolean in class Object, tree.tpe=[T0 >: ? <: ?]()Boolean
>               <tpt> //
> tree.tpe=org.apache.spark.sql.catalyst.util.ArrayData
>             )
>             Nil
>           )
>           Block( // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>             ValDef( // val x2: org.apache.spark.sql.catalyst.util.ArrayData
>               <synthetic> <triedcooking>
>               "x2"
>               <tpt> //
> tree.tpe=org.apache.spark.sql.catalyst.util.ArrayData
>               Typed( //
> tree.tpe=org.apache.spark.sql.catalyst.util.ArrayData
>                 Apply( // final def $asInstanceOf[T0 >: ? <: ?](): T0 in
> class Object, tree.tpe=org.apache.spark.sql.catalyst.util.ArrayData
>                   TypeApply( // final def $asInstanceOf[T0 >: ? <: ?]():
> T0 in class Object, tree.tpe=()org.apache.spark.sql.catalyst.util.ArrayData
>                     "x1"."$asInstanceOf" // final def $asInstanceOf[T0 >:
> ? <: ?](): T0 in class Object, tree.tpe=[T0 >: ? <: ?]()T0
>                     <tpt> //
> tree.tpe=org.apache.spark.sql.catalyst.util.ArrayData
>                   )
>                   Nil
>                 )
>                 <tpt> //
> tree.tpe=org.apache.spark.sql.catalyst.util.ArrayData
>               )
>             )
>             Apply( // case def matchEnd4(x:
> org.apache.spark.sql.test.ExamplePoint):
> org.apache.spark.sql.test.ExamplePoint,
> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>               "matchEnd4" // case def matchEnd4(x:
> org.apache.spark.sql.test.ExamplePoint):
> org.apache.spark.sql.test.ExamplePoint, tree.tpe=(x:
> org.apache.spark.sql.test.ExamplePoint)org.apache.spark.sql.test.ExamplePoint
>               Apply( // def <init>(x: Double,y: Double):
> org.apache.spark.sql.test.ExamplePoint in class ExamplePoint,
> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>                 new org.apache.spark.sql.test.ExamplePoint."<init>" // def
> <init>(x: Double,y: Double): org.apache.spark.sql.test.ExamplePoint in
> class ExamplePoint, tree.tpe=(x: Double, y:
> Double)org.apache.spark.sql.test.ExamplePoint
>                 // 2 arguments
>                 Apply( // def getDouble(x$1: Int): Double in trait
> SpecializedGetters, tree.tpe=Double
>                   "x2"."getDouble" // def getDouble(x$1: Int): Double in
> trait SpecializedGetters, tree.tpe=(x$1: Int)Double
>                   0
>                 )
>                 Apply( // def getDouble(x$1: Int): Double in trait
> SpecializedGetters, tree.tpe=Double
>                   "x2"."getDouble" // def getDouble(x$1: Int): Double in
> trait SpecializedGetters, tree.tpe=(x$1: Int)Double
>                   1
>                 )
>               )
>             )
>           )
>           Apply( // case def case6():
> org.apache.spark.sql.test.ExamplePoint,
> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>             "case6" // case def case6():
> org.apache.spark.sql.test.ExamplePoint,
> tree.tpe=()org.apache.spark.sql.test.ExamplePoint
>             Nil
>           )
>         )
>       )
>       LabelDef( // case def case6():
> org.apache.spark.sql.test.ExamplePoint,
> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>         ()
>         Apply( // case def matchEnd4(x:
> org.apache.spark.sql.test.ExamplePoint):
> org.apache.spark.sql.test.ExamplePoint,
> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>           "matchEnd4" // case def matchEnd4(x:
> org.apache.spark.sql.test.ExamplePoint):
> org.apache.spark.sql.test.ExamplePoint, tree.tpe=(x:
> org.apache.spark.sql.test.ExamplePoint)org.apache.spark.sql.test.ExamplePoint
>           Throw( // tree.tpe=Nothing
>             Apply( // def <init>(obj: Object): MatchError in class
> MatchError, tree.tpe=MatchError
>               new MatchError."<init>" // def <init>(obj: Object):
> MatchError in class MatchError, tree.tpe=(obj: Object)MatchError
>               "x1" // case val x1: Object, tree.tpe=Object
>             )
>           )
>         )
>       )
>       LabelDef( // case def matchEnd4(x:
> org.apache.spark.sql.test.ExamplePoint):
> org.apache.spark.sql.test.ExamplePoint,
> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>         "x" // x: org.apache.spark.sql.test.ExamplePoint,
> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>         "x" // x: org.apache.spark.sql.test.ExamplePoint,
> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>       )
>     )
>   )
>   DefDef( // override def userClass(): Class in class ExamplePointUDT
>     <method> override
>     "userClass"
>     []
>     List(Nil)
>     <tpt> // tree.tpe=Class
>     classOf[org.apache.spark.sql.test.ExamplePoint]
>   )
>   DefDef( // override private[package spark] def asNullable():
> org.apache.spark.sql.test.ExamplePointUDT in class ExamplePointUDT
>     <method> override <triedcooking>
>     "asNullable"
>     []
>     List(Nil)
>     <tpt> // tree.tpe=org.apache.spark.sql.test.ExamplePointUDT
>     This(<empty>)private[package sql] class ExamplePointUDT extends
> UserDefinedType in package test,
> tree.tpe=org.apache.spark.sql.test.ExamplePointUDT
>   )
>   DefDef( // override def asNullable():
> org.apache.spark.sql.types.DataType in class ExamplePointUDT
>     <method> override <bridge>
>     "asNullable"
>     []
>     List(Nil)
>     <tpt> // tree.tpe=org.apache.spark.sql.types.DataType
>     Apply( // override private[package spark] def asNullable():
> org.apache.spark.sql.test.ExamplePointUDT in class ExamplePointUDT,
> tree.tpe=org.apache.spark.sql.test.ExamplePointUDT
>       ExamplePointUDT.this."asNullable" // override private[package spark]
> def asNullable(): org.apache.spark.sql.test.ExamplePointUDT in class
> ExamplePointUDT, tree.tpe=()org.apache.spark.sql.test.ExamplePointUDT
>       Nil
>     )
>   )
>   DefDef( // override def asNullable():
> org.apache.spark.sql.types.UserDefinedType in class ExamplePointUDT
>     <method> override <bridge>
>     "asNullable"
>     []
>     List(Nil)
>     <tpt> // tree.tpe=org.apache.spark.sql.types.UserDefinedType
>     Apply( // override private[package spark] def asNullable():
> org.apache.spark.sql.test.ExamplePointUDT in class ExamplePointUDT,
> tree.tpe=org.apache.spark.sql.test.ExamplePointUDT
>       ExamplePointUDT.this."asNullable" // override private[package spark]
> def asNullable(): org.apache.spark.sql.test.ExamplePointUDT in class
> ExamplePointUDT, tree.tpe=()org.apache.spark.sql.test.ExamplePointUDT
>       Nil
>     )
>   )
>   DefDef( // override def deserialize(datum: Object): Object in class
> ExamplePointUDT
>     <method> override <bridge>
>     "deserialize"
>     []
>     // 1 parameter list
>     ValDef( // datum: Object
>       <param> <triedcooking>
>       "datum"
>       <tpt> // tree.tpe=Object
>       <empty>
>     )
>     <tpt> // tree.tpe=Object
>     Apply( // override def deserialize(datum: Object):
> org.apache.spark.sql.test.ExamplePoint in class ExamplePointUDT,
> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>       ExamplePointUDT.this."deserialize" // override def
> deserialize(datum: Object): org.apache.spark.sql.test.ExamplePoint in class
> ExamplePointUDT, tree.tpe=(datum:
> Object)org.apache.spark.sql.test.ExamplePoint
>       "datum" // datum: Object, tree.tpe=Object
>     )
>   )
>   DefDef( // override def serialize(obj: Object): Object in class
> ExamplePointUDT
>     <method> override <bridge>
>     "serialize"
>     []
>     // 1 parameter list
>     ValDef( // obj: Object
>       <param> <triedcooking>
>       "obj"
>       <tpt> // tree.tpe=Object
>       <empty>
>     )
>     <tpt> // tree.tpe=Object
>     Apply( // override def serialize(obj: Object):
> org.apache.spark.sql.catalyst.util.GenericArrayData in class
> ExamplePointUDT,
> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>       ExamplePointUDT.this."serialize" // override def serialize(obj:
> Object): org.apache.spark.sql.catalyst.util.GenericArrayData in class
> ExamplePointUDT, tree.tpe=(obj:
> Object)org.apache.spark.sql.catalyst.util.GenericArrayData
>       "obj" // obj: Object, tree.tpe=Object
>     )
>   )
>   DefDef( // def <init>(): org.apache.spark.sql.test.ExamplePointUDT in
> class ExamplePointUDT
>     <method>
>     "<init>"
>     []
>     List(Nil)
>     <tpt> // tree.tpe=org.apache.spark.sql.test.ExamplePointUDT
>     Block( // tree.tpe=Unit
>       Apply( // def <init>(): org.apache.spark.sql.types.UserDefinedType
> in class UserDefinedType,
> tree.tpe=org.apache.spark.sql.types.UserDefinedType
>         ExamplePointUDT.super."<init>" // def <init>():
> org.apache.spark.sql.types.UserDefinedType in class UserDefinedType,
> tree.tpe=()org.apache.spark.sql.types.UserDefinedType
>         Nil
>       )
>       ()
>     )
>   )
> )
> == Expanded type of tree ==
> *ConstantType(*
> *  value = Constant(org.apache.spark.sql.test.ExamplePoint)*
> *)*
> *uncaught exception during compilation: java.lang.AssertionError*
>
> *Error:scala: Error: assertion failed: List(object package$DebugNode,
> object package$DebugNode)*
> *java.lang.AssertionError: assertion failed: List(object
> package$DebugNode, object package$DebugNode)*
> at scala.reflect.internal.Symbols$Symbol.suchThat(Symbols.scala:1678)
> at
> scala.reflect.internal.Symbols$ClassSymbol.companionModule0(Symbols.scala:2988)
> at
> scala.reflect.internal.Symbols$ClassSymbol.companionModule(Symbols.scala:2991)
> at
> scala.tools.nsc.backend.jvm.GenASM$JPlainBuilder.genClass(GenASM.scala:1371)
> at scala.tools.nsc.backend.jvm.GenASM$AsmPhase.run(GenASM.scala:120)
> at scala.tools.nsc.Global$Run.compileUnitsInternal(Global.scala:1583)
> at scala.tools.nsc.Global$Run.compileUnits(Global.scala:1557)
> at scala.tools.nsc.Global$Run.compileSources(Global.scala:1553)
> at scala.tools.nsc.Global$Run.compile(Global.scala:1662)
> at xsbt.CachedCompiler0.run(CompilerInterface.scala:126)
> at xsbt.CachedCompiler0.run(CompilerInterface.scala:102)
> at xsbt.CompilerInterface.run(CompilerInterface.scala:27)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at sbt.compiler.AnalyzingCompiler.call(AnalyzingCompiler.scala:102)
> at sbt.compiler.AnalyzingCompiler.compile(AnalyzingCompiler.scala:48)
> at sbt.compiler.AnalyzingCompiler.compile(AnalyzingCompiler.scala:41)
> at
> sbt.compiler.AggressiveCompile$$anonfun$6$$anonfun$compileScala$1$1$$anonfun$apply$3$$anonfun$apply$1.apply$mcV$sp(AggressiveCompile.scala:106)
> at
> sbt.compiler.AggressiveCompile$$anonfun$6$$anonfun$compileScala$1$1$$anonfun$apply$3$$anonfun$apply$1.apply(AggressiveCompile.scala:106)
> at
> sbt.compiler.AggressiveCompile$$anonfun$6$$anonfun$compileScala$1$1$$anonfun$apply$3$$anonfun$apply$1.apply(AggressiveCompile.scala:106)
> at
> sbt.compiler.AggressiveCompile.sbt$compiler$AggressiveCompile$$timed(AggressiveCompile.scala:179)
> at
> sbt.compiler.AggressiveCompile$$anonfun$6$$anonfun$compileScala$1$1$$anonfun$apply$3.apply(AggressiveCompile.scala:105)
> at
> sbt.compiler.AggressiveCompile$$anonfun$6$$anonfun$compileScala$1$1$$anonfun$apply$3.apply(AggressiveCompile.scala:102)
> at scala.Option.foreach(Option.scala:245)
> at
> sbt.compiler.AggressiveCompile$$anonfun$6$$anonfun$compileScala$1$1.apply(AggressiveCompile.scala:102)
> at
> sbt.compiler.AggressiveCompile$$anonfun$6$$anonfun$compileScala$1$1.apply(AggressiveCompile.scala:102)
> at scala.Option.foreach(Option.scala:245)
> at
> sbt.compiler.AggressiveCompile$$anonfun$6.compileScala$1(AggressiveCompile.scala:102)
> at
> sbt.compiler.AggressiveCompile$$anonfun$6.apply(AggressiveCompile.scala:151)
> at
> sbt.compiler.AggressiveCompile$$anonfun$6.apply(AggressiveCompile.scala:89)
> at sbt.inc.IncrementalCompile$$anonfun$doCompile$1.apply(Compile.scala:40)
> at sbt.inc.IncrementalCompile$$anonfun$doCompile$1.apply(Compile.scala:38)
> at sbt.inc.IncrementalCommon.cycle(Incremental.scala:103)
> at sbt.inc.Incremental$$anonfun$1.apply(Incremental.scala:39)
> at sbt.inc.Incremental$$anonfun$1.apply(Incremental.scala:38)
> at sbt.inc.Incremental$.manageClassfiles(Incremental.scala:69)
> at sbt.inc.Incremental$.compile(Incremental.scala:38)
> at sbt.inc.IncrementalCompile$.apply(Compile.scala:28)
> at sbt.compiler.AggressiveCompile.compile2(AggressiveCompile.scala:170)
> at sbt.compiler.AggressiveCompile.compile1(AggressiveCompile.scala:73)
> at
> org.jetbrains.jps.incremental.scala.local.SbtCompiler.compile(SbtCompiler.scala:66)
> at
> org.jetbrains.jps.incremental.scala.local.LocalServer.compile(LocalServer.scala:26)
> at org.jetbrains.jps.incremental.scala.remote.Main$.make(Main.scala:62)
> at org.jetbrains.jps.incremental.scala.remote.Main$.nailMain(Main.scala:20)
> at org.jetbrains.jps.incremental.scala.remote.Main.nailMain(Main.scala)
> at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:606)
> at com.martiansoftware.nailgun.NGSession.run(NGSession.java:319)
>
> I just highlighted some error message that I think important as *bold and
> red.*
>
> This really bothered me for several days, I don't know how to get through.
> Any suggestions? Thanks.
>

Reply via email to