This also bother me for a long time. I suspect the intellij builder
conflicts with the sbt/maven builder.

I resolve this issue by rebuild spark in intellij.  You may meet
compilation issue when building it in intellij.
For that you need to put external/flume-sink/target/java on the source
build path.



On Wed, Nov 18, 2015 at 12:02 PM, Ted Yu <yuzhih...@gmail.com> wrote:

> Is the Scala version in Intellij the same as the one used by sbt ?
>
> Cheers
>
> On Tue, Nov 17, 2015 at 6:45 PM, 金国栋 <scu...@gmail.com> wrote:
>
>> Hi!
>>
>> I tried to build spark source code from github, and I successfully built
>> it from command line using `*sbt/sbt assembly*`. While I encountered an
>> error when compiling the project in Intellij IDEA(V14.1.5).
>>
>>
>> The error log is below:
>> *Error:scala: *
>> *     while compiling:
>> /Users/ray/Documents/P01_Project/Spark-Github/spark/sql/core/src/main/scala/org/apache/spark/sql/util/QueryExecutionListener.scala*
>> *        during phase: jvm*
>>      library version: version 2.10.5
>>     compiler version: version 2.10.5
>>   reconstructed args: -nobootcp -javabootclasspath : -deprecation
>> -feature -classpath
>> /Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/lib/ant-javafx.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/lib/dt.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/lib/javafx-mx.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/lib/jconsole.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/lib/sa-jdi.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/lib/tools.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/charsets.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/deploy.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/htmlconverter.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/javaws.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/jce.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/jfr.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/jfxswt.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/jsse.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/management-agent.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/plugin.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/resources.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/rt.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/cldrdata.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/dnsns.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/jfxrt.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/localedata.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/nashorn.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/sunec.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/sunjce_provider.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/sunpkcs11.jar:/Library/Java/JavaVirtualMachines/jdk1.8.0_05.jdk/Contents/Home/jre/lib/ext/zipfs.jar:/Users/ray/Documents/P01_Project/Spark-Github/spark/sql/core/target/scala-2.10/classes:/Users/ray/Documents/P01_Project/Spark-Github/spark/core/target/scala-2.10/classes:/Users/ray/.m2/repository/org/apache/avro/avro-mapred/1.7.7/avro-mapred-1.7.7-hadoop2.jar:/Users/ray/.m2/repository/org/apache/avro/avro-ipc/1.7.7/avro-ipc-1.7.7.jar:/Users/ray/.m2/repository/org/apache/avro/avro-ipc/1.7.7/avro-ipc-1.7.7-tests.jar:/Users/ray/.m2/repository/com/twitter/chill_2.10/0.5.0/chill_2.10-0.5.0.jar:/Users/ray/.m2/repository/com/esotericsoftware/kryo/kryo/2.21/kryo-2.21.jar:/Users/ray/.m2/repository/com/esotericsoftware/reflectasm/reflectasm/1.07/reflectasm-1.07-shaded.jar:/Users/ray/.m2/repository/com/esotericsoftware/minlog/minlog/1.2/minlog-1.2.jar:/Users/ray/.m2/repository/com/twitter/chill-java/0.5.0/chill-java-0.5.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-client/2.2.0/hadoop-client-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-common/2.2.0/hadoop-common-2.2.0.jar:/Users/ray/.m2/repository/commons-cli/commons-cli/1.2/commons-cli-1.2.jar:/Users/ray/.m2/repository/org/apache/commons/commons-math/2.1/commons-math-2.1.jar:/Users/ray/.m2/repository/xmlenc/xmlenc/0.52/xmlenc-0.52.jar:/Users/ray/.m2/repository/commons-configuration/commons-configuration/1.6/commons-configuration-1.6.jar:/Users/ray/.m2/repository/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar:/Users/ray/.m2/repository/commons-digester/commons-digester/1.8/commons-digester-1.8.jar:/Users/ray/.m2/repository/commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar:/Users/ray/.m2/repository/commons-beanutils/commons-beanutils-core/1.8.0/commons-beanutils-core-1.8.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-auth/2.2.0/hadoop-auth-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-hdfs/2.2.0/hadoop-hdfs-2.2.0.jar:/Users/ray/.m2/repository/org/mortbay/jetty/jetty-util/6.1.26/jetty-util-6.1.26.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-mapreduce-client-app/2.2.0/hadoop-mapreduce-client-app-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-mapreduce-client-common/2.2.0/hadoop-mapreduce-client-common-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-yarn-client/2.2.0/hadoop-yarn-client-2.2.0.jar:/Users/ray/.m2/repository/com/google/inject/guice/3.0/guice-3.0.jar:/Users/ray/.m2/repository/javax/inject/javax.inject/1/javax.inject-1.jar:/Users/ray/.m2/repository/aopalliance/aopalliance/1.0/aopalliance-1.0.jar:/Users/ray/.m2/repository/com/sun/jersey/jersey-test-framework/jersey-test-framework-grizzly2/1.9/jersey-test-framework-grizzly2-1.9.jar:/Users/ray/.m2/repository/com/sun/jersey/jersey-test-framework/jersey-test-framework-core/1.9/jersey-test-framework-core-1.9.jar:/Users/ray/.m2/repository/javax/servlet/javax.servlet-api/3.0.1/javax.servlet-api-3.0.1.jar:/Users/ray/.m2/repository/com/sun/jersey/jersey-client/1.9/jersey-client-1.9.jar:/Users/ray/.m2/repository/com/sun/jersey/jersey-grizzly2/1.9/jersey-grizzly2-1.9.jar:/Users/ray/.m2/repository/org/glassfish/grizzly/grizzly-http/2.1.2/grizzly-http-2.1.2.jar:/Users/ray/.m2/repository/org/glassfish/grizzly/grizzly-framework/2.1.2/grizzly-framework-2.1.2.jar:/Users/ray/.m2/repository/org/glassfish/gmbal/gmbal-api-only/3.0.0-b023/gmbal-api-only-3.0.0-b023.jar:/Users/ray/.m2/repository/org/glassfish/external/management-api/3.0.0-b012/management-api-3.0.0-b012.jar:/Users/ray/.m2/repository/org/glassfish/grizzly/grizzly-http-server/2.1.2/grizzly-http-server-2.1.2.jar:/Users/ray/.m2/repository/org/glassfish/grizzly/grizzly-rcm/2.1.2/grizzly-rcm-2.1.2.jar:/Users/ray/.m2/repository/org/glassfish/grizzly/grizzly-http-servlet/2.1.2/grizzly-http-servlet-2.1.2.jar:/Users/ray/.m2/repository/org/glassfish/javax.servlet/3.1/javax.servlet-3.1.jar:/Users/ray/.m2/repository/com/sun/jersey/jersey-json/1.9/jersey-json-1.9.jar:/Users/ray/.m2/repository/org/codehaus/jettison/jettison/1.1/jettison-1.1.jar:/Users/ray/.m2/repository/com/sun/xml/bind/jaxb-impl/2.2.3-1/jaxb-impl-2.2.3-1.jar:/Users/ray/.m2/repository/javax/xml/bind/jaxb-api/2.2.2/jaxb-api-2.2.2.jar:/Users/ray/.m2/repository/javax/activation/activation/1.1/activation-1.1.jar:/Users/ray/.m2/repository/org/codehaus/jackson/jackson-jaxrs/1.9.13/jackson-jaxrs-1.9.13.jar:/Users/ray/.m2/repository/org/codehaus/jackson/jackson-xc/1.9.13/jackson-xc-1.9.13.jar:/Users/ray/.m2/repository/com/sun/jersey/contribs/jersey-guice/1.9/jersey-guice-1.9.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-yarn-server-common/2.2.0/hadoop-yarn-server-common-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-mapreduce-client-shuffle/2.2.0/hadoop-mapreduce-client-shuffle-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-yarn-api/2.2.0/hadoop-yarn-api-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-mapreduce-client-core/2.2.0/hadoop-mapreduce-client-core-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-yarn-common/2.2.0/hadoop-yarn-common-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-mapreduce-client-jobclient/2.2.0/hadoop-mapreduce-client-jobclient-2.2.0.jar:/Users/ray/.m2/repository/org/apache/hadoop/hadoop-annotations/2.2.0/hadoop-annotations-2.2.0.jar:/Users/ray/Documents/P01_Project/Spark-Github/spark/launcher/target/scala-2.10/classes:/Users/ray/Documents/P01_Project/Spark-Github/spark/network/common/target/scala-2.10/classes:/Users/ray/.m2/repository/com/google/guava/guava/14.0.1/guava-14.0.1.jar:/Users/ray/Documents/P01_Project/Spark-Github/spark/network/shuffle/target/scala-2.10/classes:/Users/ray/.m2/repository/org/fusesource/leveldbjni/leveldbjni-all/1.8/leveldbjni-all-1.8.jar:/Users/ray/Documents/P01_Project/Spark-Github/spark/unsafe/target/scala-2.10/classes:/Users/ray/.m2/repository/net/java/dev/jets3t/jets3t/0.7.1/jets3t-0.7.1.jar:/Users/ray/.m2/repository/commons-httpclient/commons-httpclient/3.1/commons-httpclient-3.1.jar:/Users/ray/.m2/repository/org/apache/curator/curator-recipes/2.4.0/curator-recipes-2.4.0.jar:/Users/ray/.m2/repository/org/apache/curator/curator-framework/2.4.0/curator-framework-2.4.0.jar:/Users/ray/.m2/repository/org/apache/curator/curator-client/2.4.0/curator-client-2.4.0.jar:/Users/ray/.m2/repository/org/apache/zookeeper/zookeeper/3.4.5/zookeeper-3.4.5.jar:/Users/ray/.m2/repository/jline/jline/0.9.94/jline-0.9.94.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-plus/8.1.14.v20131031/jetty-plus-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/orbit/javax.transaction/1.1.1.v201105210645/javax.transaction-1.1.1.v201105210645.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-webapp/8.1.14.v20131031/jetty-webapp-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-xml/8.1.14.v20131031/jetty-xml-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-jndi/8.1.14.v20131031/jetty-jndi-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/orbit/javax.mail.glassfish/1.4.1.v201005082020/javax.mail.glassfish-1.4.1.v201005082020.jar:/Users/ray/.m2/repository/org/eclipse/jetty/orbit/javax.activation/1.1.0.v201105071233/javax.activation-1.1.0.v201105071233.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-security/8.1.14.v20131031/jetty-security-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-util/8.1.14.v20131031/jetty-util-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-server/8.1.14.v20131031/jetty-server-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-http/8.1.14.v20131031/jetty-http-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-io/8.1.14.v20131031/jetty-io-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-continuation/8.1.14.v20131031/jetty-continuation-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/jetty-servlet/8.1.14.v20131031/jetty-servlet-8.1.14.v20131031.jar:/Users/ray/.m2/repository/org/eclipse/jetty/orbit/javax.servlet/3.0.0.v201112011016/javax.servlet-3.0.0.v201112011016.jar:/Users/ray/.m2/repository/org/apache/commons/commons-lang3/3.3.2/commons-lang3-3.3.2.jar:/Users/ray/.m2/repository/org/apache/commons/commons-math3/3.4.1/commons-math3-3.4.1.jar:/Users/ray/.m2/repository/com/google/code/findbugs/jsr305/1.3.9/jsr305-1.3.9.jar:/Users/ray/.m2/repository/org/slf4j/slf4j-api/1.7.10/slf4j-api-1.7.10.jar:/Users/ray/.m2/repository/org/slf4j/jul-to-slf4j/1.7.10/jul-to-slf4j-1.7.10.jar:/Users/ray/.m2/repository/org/slf4j/jcl-over-slf4j/1.7.10/jcl-over-slf4j-1.7.10.jar:/Users/ray/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar:/Users/ray/.m2/repository/org/slf4j/slf4j-log4j12/1.7.10/slf4j-log4j12-1.7.10.jar:/Users/ray/.m2/repository/com/ning/compress-lzf/1.0.3/compress-lzf-1.0.3.jar:/Users/ray/.m2/repository/org/xerial/snappy/snappy-java/1.1.2/snappy-java-1.1.2.jar:/Users/ray/.m2/repository/net/jpountz/lz4/lz4/1.3.0/lz4-1.3.0.jar:/Users/ray/.m2/repository/org/roaringbitmap/RoaringBitmap/0.4.5/RoaringBitmap-0.4.5.jar:/Users/ray/.m2/repository/commons-net/commons-net/2.2/commons-net-2.2.jar:/Users/ray/.m2/repository/com/typesafe/akka/akka-remote_2.10/2.3.11/akka-remote_2.10-2.3.11.jar:/Users/ray/.m2/repository/com/typesafe/akka/akka-actor_2.10/2.3.11/akka-actor_2.10-2.3.11.jar:/Users/ray/.m2/repository/com/typesafe/config/1.2.1/config-1.2.1.jar:/Users/ray/.m2/repository/io/netty/netty/3.8.0.Final/netty-3.8.0.Final.jar:/Users/ray/.m2/repository/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar:/Users/ray/.m2/repository/org/uncommons/maths/uncommons-maths/1.2.2a/uncommons-maths-1.2.2a.jar:/Users/ray/.m2/repository/com/typesafe/akka/akka-slf4j_2.10/2.3.11/akka-slf4j_2.10-2.3.11.jar:/Users/ray/.m2/repository/org/scala-lang/scala-library/2.10.5/scala-library-2.10.5.jar:/Users/ray/.m2/repository/org/json4s/json4s-jackson_2.10/3.2.10/json4s-jackson_2.10-3.2.10.jar:/Users/ray/.m2/repository/org/json4s/json4s-core_2.10/3.2.10/json4s-core_2.10-3.2.10.jar:/Users/ray/.m2/repository/org/json4s/json4s-ast_2.10/3.2.10/json4s-ast_2.10-3.2.10.jar:/Users/ray/.m2/repository/org/scala-lang/scalap/2.10.5/scalap-2.10.5.jar:/Users/ray/.m2/repository/org/scala-lang/scala-compiler/2.10.5/scala-compiler-2.10.5.jar:/Users/ray/.m2/repository/com/sun/jersey/jersey-server/1.9/jersey-server-1.9.jar:/Users/ray/.m2/repository/asm/asm/3.1/asm-3.1.jar:/Users/ray/.m2/repository/com/sun/jersey/jersey-core/1.9/jersey-core-1.9.jar:/Users/ray/.m2/repository/org/apache/mesos/mesos/0.21.1/mesos-0.21.1-shaded-protobuf.jar:/Users/ray/.m2/repository/io/netty/netty-all/4.0.29.Final/netty-all-4.0.29.Final.jar:/Users/ray/.m2/repository/com/clearspring/analytics/stream/2.7.0/stream-2.7.0.jar:/Users/ray/.m2/repository/io/dropwizard/metrics/metrics-core/3.1.2/metrics-core-3.1.2.jar:/Users/ray/.m2/repository/io/dropwizard/metrics/metrics-jvm/3.1.2/metrics-jvm-3.1.2.jar:/Users/ray/.m2/repository/io/dropwizard/metrics/metrics-json/3.1.2/metrics-json-3.1.2.jar:/Users/ray/.m2/repository/io/dropwizard/metrics/metrics-graphite/3.1.2/metrics-graphite-3.1.2.jar:/Users/ray/.m2/repository/com/fasterxml/jackson/module/jackson-module-scala_2.10/2.4.4/jackson-module-scala_2.10-2.4.4.jar:/Users/ray/.m2/repository/com/thoughtworks/paranamer/paranamer/2.6/paranamer-2.6.jar:/Users/ray/.m2/repository/org/apache/ivy/ivy/2.4.0/ivy-2.4.0.jar:/Users/ray/.m2/repository/oro/oro/2.0.8/oro-2.0.8.jar:/Users/ray/.m2/repository/org/tachyonproject/tachyon-client/0.8.1/tachyon-client-0.8.1.jar:/Users/ray/.m2/repository/commons-lang/commons-lang/2.6/commons-lang-2.6.jar:/Users/ray/.m2/repository/commons-io/commons-io/2.4/commons-io-2.4.jar:/Users/ray/.m2/repository/org/tachyonproject/tachyon-underfs-hdfs/0.8.1/tachyon-underfs-hdfs-0.8.1.jar:/Users/ray/.m2/repository/org/tachyonproject/tachyon-underfs-s3/0.8.1/tachyon-underfs-s3-0.8.1.jar:/Users/ray/.m2/repository/org/tachyonproject/tachyon-underfs-local/0.8.1/tachyon-underfs-local-0.8.1.jar:/Users/ray/.m2/repository/net/razorvine/pyrolite/4.9/pyrolite-4.9.jar:/Users/ray/.m2/repository/net/sf/py4j/py4j/0.9/py4j-0.9.jar:/Users/ray/Documents/P01_Project/Spark-Github/spark/sql/catalyst/target/scala-2.10/classes:/Users/ray/.m2/repository/org/scala-lang/scala-reflect/2.10.5/scala-reflect-2.10.5.jar:/Users/ray/.m2/repository/org/codehaus/janino/janino/2.7.8/janino-2.7.8.jar:/Users/ray/.m2/repository/org/codehaus/janino/commons-compiler/2.7.8/commons-compiler-2.7.8.jar:/Users/ray/.m2/repository/org/apache/parquet/parquet-column/1.7.0/parquet-column-1.7.0.jar:/Users/ray/.m2/repository/org/apache/parquet/parquet-common/1.7.0/parquet-common-1.7.0.jar:/Users/ray/.m2/repository/org/apache/parquet/parquet-encoding/1.7.0/parquet-encoding-1.7.0.jar:/Users/ray/.m2/repository/org/apache/parquet/parquet-generator/1.7.0/parquet-generator-1.7.0.jar:/Users/ray/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar:/Users/ray/.m2/repository/org/apache/parquet/parquet-hadoop/1.7.0/parquet-hadoop-1.7.0.jar:/Users/ray/.m2/repository/org/apache/parquet/parquet-format/2.3.0-incubating/parquet-format-2.3.0-incubating.jar:/Users/ray/.m2/repository/org/apache/parquet/parquet-jackson/1.7.0/parquet-jackson-1.7.0.jar:/Users/ray/.m2/repository/org/codehaus/jackson/jackson-mapper-asl/1.9.13/jackson-mapper-asl-1.9.13.jar:/Users/ray/.m2/repository/org/codehaus/jackson/jackson-core-asl/1.9.13/jackson-core-asl-1.9.13.jar:/Users/ray/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.4.4/jackson-databind-2.4.4.jar:/Users/ray/.m2/repository/com/fasterxml/jackson/core/jackson-annotations/2.4.4/jackson-annotations-2.4.4.jar:/Users/ray/.m2/repository/com/fasterxml/jackson/core/jackson-core/2.4.4/jackson-core-2.4.4.jar:/Users/ray/.m2/repository/org/apache/avro/avro/1.7.7/avro-1.7.7.jar:/Users/ray/.m2/repository/org/apache/commons/commons-compress/1.4.1/commons-compress-1.4.1.jar:/Users/ray/.m2/repository/org/tukaani/xz/1.0/xz-1.0.jar:/Users/ray/.m2/repository/org/objenesis/objenesis/1.0/objenesis-1.0.jar:/Users/ray/.m2/repository/org/spark-project/spark/unused/1.0.0/unused-1.0.0.jar
>> -unchecked
>>   last tree to typer:
>> Literal(Constant(org.apache.spark.sql.test.ExamplePoint))
>>               symbol: null
>>    symbol definition: null
>>                  tpe:
>> Class(classOf[org.apache.spark.sql.test.ExamplePoint])
>>        symbol owners:
>>       context owners: class ExamplePointUDT -> package test
>> == Enclosing template or block ==
>> Template( // val <local ExamplePointUDT>: <notype> in class
>> ExamplePointUDT, tree.tpe=org.apache.spark.sql.test.ExamplePointUDT
>>   "org.apache.spark.sql.types.UserDefinedType" // parents
>>   ValDef(
>>     private
>>     "_"
>>     <tpt>
>>     <empty>
>>   )
>>   // 11 statements
>>   DefDef( // override def sqlType(): org.apache.spark.sql.types.DataType
>> in class ExamplePointUDT
>>     <method> override
>>     "sqlType"
>>     []
>>     List(Nil)
>>     <tpt> // tree.tpe=org.apache.spark.sql.types.DataType
>>     Apply( // def <init>(elementType:
>> org.apache.spark.sql.types.DataType,containsNull: Boolean):
>> org.apache.spark.sql.types.ArrayType in class ArrayType,
>> tree.tpe=org.apache.spark.sql.types.ArrayType
>>       new org.apache.spark.sql.types.ArrayType."<init>" // def
>> <init>(elementType: org.apache.spark.sql.types.DataType,containsNull:
>> Boolean): org.apache.spark.sql.types.ArrayType in class ArrayType,
>> tree.tpe=(elementType: org.apache.spark.sql.types.DataType, containsNull:
>> Boolean)org.apache.spark.sql.types.ArrayType
>>       // 2 arguments
>>       "org"."apache"."spark"."sql"."types"."DoubleType" // case object
>> DoubleType in package types,
>> tree.tpe=org.apache.spark.sql.types.DoubleType.type
>>       false
>>     )
>>   )
>>   DefDef( // override def pyUDT(): String in class ExamplePointUDT
>>     <method> override
>>     "pyUDT"
>>     []
>>     List(Nil)
>>     <tpt> // tree.tpe=String
>>     "pyspark.sql.tests.ExamplePointUDT"
>>   )
>>   DefDef( // override def serialize(obj: Object):
>> org.apache.spark.sql.catalyst.util.GenericArrayData in class ExamplePointUDT
>>     <method> override <triedcooking>
>>     "serialize"
>>     []
>>     // 1 parameter list
>>     ValDef( // obj: Object
>>       <param> <triedcooking>
>>       "obj"
>>       <tpt> // tree.tpe=Object
>>       <empty>
>>     )
>>     <tpt> // tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>>     Block( // tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>>       // 3 statements
>>       ValDef( // case val x1: Object
>>         case <synthetic> <triedcooking>
>>         "x1"
>>         <tpt> // tree.tpe=Object
>>         "obj" // obj: Object, tree.tpe=Object
>>       )
>>       LabelDef( // case def case5():
>> org.apache.spark.sql.catalyst.util.GenericArrayData,
>> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>>         ()
>>         If( //
>> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>>           Apply( // final def $isInstanceOf[T0 >: ? <: ?](): Boolean in
>> class Object, tree.tpe=Boolean
>>             TypeApply( // final def $isInstanceOf[T0 >: ? <: ?]():
>> Boolean in class Object, tree.tpe=()Boolean
>>               "x1"."$isInstanceOf" // final def $isInstanceOf[T0 >: ? <:
>> ?](): Boolean in class Object, tree.tpe=[T0 >: ? <: ?]()Boolean
>>               <tpt> // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>             )
>>             Nil
>>           )
>>           Block( //
>> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>>             ValDef( // val x2: org.apache.spark.sql.test.ExamplePoint
>>               <synthetic> <triedcooking>
>>               "x2"
>>               <tpt> // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>               Typed( // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>                 Apply( // final def $asInstanceOf[T0 >: ? <: ?](): T0 in
>> class Object, tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>                   TypeApply( // final def $asInstanceOf[T0 >: ? <: ?]():
>> T0 in class Object, tree.tpe=()org.apache.spark.sql.test.ExamplePoint
>>                     "x1"."$asInstanceOf" // final def $asInstanceOf[T0 >:
>> ? <: ?](): T0 in class Object, tree.tpe=[T0 >: ? <: ?]()T0
>>                     <tpt> //
>> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>                   )
>>                   Nil
>>                 )
>>                 <tpt> // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>               )
>>             )
>>             Apply( // case def matchEnd4(x:
>> org.apache.spark.sql.catalyst.util.GenericArrayData):
>> org.apache.spark.sql.catalyst.util.GenericArrayData,
>> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>>               "matchEnd4" // case def matchEnd4(x:
>> org.apache.spark.sql.catalyst.util.GenericArrayData):
>> org.apache.spark.sql.catalyst.util.GenericArrayData, tree.tpe=(x:
>> org.apache.spark.sql.catalyst.util.GenericArrayData)org.apache.spark.sql.catalyst.util.GenericArrayData
>>               Block( //
>> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>>                 // 3 statements
>>                 ValDef( // val output: Array[Object]
>>                   <triedcooking>
>>                   "output"
>>                   <tpt> // tree.tpe=Array[Object]
>>                   Apply( // def <init>(_length: Int): Array[T] in class
>> Array, tree.tpe=Array[Object]
>>                     new Array[Object]."<init>" // def <init>(_length:
>> Int): Array[T] in class Array, tree.tpe=(_length: Int)Array[Object]
>>                     2
>>                   )
>>                 )
>>                 Apply( // def update(i: Int,x: T): Unit in class Array,
>> tree.tpe=Unit
>>                   "output"."update" // def update(i: Int,x: T): Unit in
>> class Array, tree.tpe=(i: Int, x: Object)Unit
>>                   // 2 arguments
>>                   0
>>                   Apply( // def box(x: Double): Double in object Double,
>> tree.tpe=Object
>>                     "scala"."Double"."box" // def box(x: Double): Double
>> in object Double, tree.tpe=(x: Double)Double
>>                     Apply( // val x(): Double in class ExamplePoint,
>> tree.tpe=Double
>>                       "x2"."x" // val x(): Double in class ExamplePoint,
>> tree.tpe=()Double
>>                       Nil
>>                     )
>>                   )
>>                 )
>>                 Apply( // def update(i: Int,x: T): Unit in class Array,
>> tree.tpe=Unit
>>                   "output"."update" // def update(i: Int,x: T): Unit in
>> class Array, tree.tpe=(i: Int, x: Object)Unit
>>                   // 2 arguments
>>                   1
>>                   Apply( // def box(x: Double): Double in object Double,
>> tree.tpe=Object
>>                     "scala"."Double"."box" // def box(x: Double): Double
>> in object Double, tree.tpe=(x: Double)Double
>>                     Apply( // val y(): Double in class ExamplePoint,
>> tree.tpe=Double
>>                       "x2"."y" // val y(): Double in class ExamplePoint,
>> tree.tpe=()Double
>>                       Nil
>>                     )
>>                   )
>>                 )
>>                 Apply( // def <init>(array: Array[Object]):
>> org.apache.spark.sql.catalyst.util.GenericArrayData in class
>> GenericArrayData,
>> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>>                   new
>> org.apache.spark.sql.catalyst.util.GenericArrayData."<init>" // def
>> <init>(array: Array[Object]):
>> org.apache.spark.sql.catalyst.util.GenericArrayData in class
>> GenericArrayData, tree.tpe=(array:
>> Array[Object])org.apache.spark.sql.catalyst.util.GenericArrayData
>>                   "output" // val output: Array[Object],
>> tree.tpe=Array[Object]
>>                 )
>>               )
>>             )
>>           )
>>           Apply( // case def case6():
>> org.apache.spark.sql.catalyst.util.GenericArrayData,
>> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>>             "case6" // case def case6():
>> org.apache.spark.sql.catalyst.util.GenericArrayData,
>> tree.tpe=()org.apache.spark.sql.catalyst.util.GenericArrayData
>>             Nil
>>           )
>>         )
>>       )
>>       LabelDef( // case def case6():
>> org.apache.spark.sql.catalyst.util.GenericArrayData,
>> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>>         ()
>>         Apply( // case def matchEnd4(x:
>> org.apache.spark.sql.catalyst.util.GenericArrayData):
>> org.apache.spark.sql.catalyst.util.GenericArrayData,
>> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>>           "matchEnd4" // case def matchEnd4(x:
>> org.apache.spark.sql.catalyst.util.GenericArrayData):
>> org.apache.spark.sql.catalyst.util.GenericArrayData, tree.tpe=(x:
>> org.apache.spark.sql.catalyst.util.GenericArrayData)org.apache.spark.sql.catalyst.util.GenericArrayData
>>           Throw( // tree.tpe=Nothing
>>             Apply( // def <init>(obj: Object): MatchError in class
>> MatchError, tree.tpe=MatchError
>>               new MatchError."<init>" // def <init>(obj: Object):
>> MatchError in class MatchError, tree.tpe=(obj: Object)MatchError
>>               "x1" // case val x1: Object, tree.tpe=Object
>>             )
>>           )
>>         )
>>       )
>>       LabelDef( // case def matchEnd4(x:
>> org.apache.spark.sql.catalyst.util.GenericArrayData):
>> org.apache.spark.sql.catalyst.util.GenericArrayData,
>> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>>         "x" // x: org.apache.spark.sql.catalyst.util.GenericArrayData,
>> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>>         "x" // x: org.apache.spark.sql.catalyst.util.GenericArrayData,
>> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>>       )
>>     )
>>   )
>>   DefDef( // override def deserialize(datum: Object):
>> org.apache.spark.sql.test.ExamplePoint in class ExamplePointUDT
>>     <method> override <triedcooking>
>>     "deserialize"
>>     []
>>     // 1 parameter list
>>     ValDef( // datum: Object
>>       <param> <triedcooking>
>>       "datum"
>>       <tpt> // tree.tpe=Object
>>       <empty>
>>     )
>>     <tpt> // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>     Block( // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>       // 3 statements
>>       ValDef( // case val x1: Object
>>         case <synthetic> <triedcooking>
>>         "x1"
>>         <tpt> // tree.tpe=Object
>>         "datum" // datum: Object, tree.tpe=Object
>>       )
>>       LabelDef( // case def case5():
>> org.apache.spark.sql.test.ExamplePoint,
>> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>         ()
>>         If( // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>           Apply( // final def $isInstanceOf[T0 >: ? <: ?](): Boolean in
>> class Object, tree.tpe=Boolean
>>             TypeApply( // final def $isInstanceOf[T0 >: ? <: ?]():
>> Boolean in class Object, tree.tpe=()Boolean
>>               "x1"."$isInstanceOf" // final def $isInstanceOf[T0 >: ? <:
>> ?](): Boolean in class Object, tree.tpe=[T0 >: ? <: ?]()Boolean
>>               <tpt> //
>> tree.tpe=org.apache.spark.sql.catalyst.util.ArrayData
>>             )
>>             Nil
>>           )
>>           Block( // tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>             ValDef( // val x2:
>> org.apache.spark.sql.catalyst.util.ArrayData
>>               <synthetic> <triedcooking>
>>               "x2"
>>               <tpt> //
>> tree.tpe=org.apache.spark.sql.catalyst.util.ArrayData
>>               Typed( //
>> tree.tpe=org.apache.spark.sql.catalyst.util.ArrayData
>>                 Apply( // final def $asInstanceOf[T0 >: ? <: ?](): T0 in
>> class Object, tree.tpe=org.apache.spark.sql.catalyst.util.ArrayData
>>                   TypeApply( // final def $asInstanceOf[T0 >: ? <: ?]():
>> T0 in class Object, tree.tpe=()org.apache.spark.sql.catalyst.util.ArrayData
>>                     "x1"."$asInstanceOf" // final def $asInstanceOf[T0 >:
>> ? <: ?](): T0 in class Object, tree.tpe=[T0 >: ? <: ?]()T0
>>                     <tpt> //
>> tree.tpe=org.apache.spark.sql.catalyst.util.ArrayData
>>                   )
>>                   Nil
>>                 )
>>                 <tpt> //
>> tree.tpe=org.apache.spark.sql.catalyst.util.ArrayData
>>               )
>>             )
>>             Apply( // case def matchEnd4(x:
>> org.apache.spark.sql.test.ExamplePoint):
>> org.apache.spark.sql.test.ExamplePoint,
>> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>               "matchEnd4" // case def matchEnd4(x:
>> org.apache.spark.sql.test.ExamplePoint):
>> org.apache.spark.sql.test.ExamplePoint, tree.tpe=(x:
>> org.apache.spark.sql.test.ExamplePoint)org.apache.spark.sql.test.ExamplePoint
>>               Apply( // def <init>(x: Double,y: Double):
>> org.apache.spark.sql.test.ExamplePoint in class ExamplePoint,
>> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>                 new org.apache.spark.sql.test.ExamplePoint."<init>" //
>> def <init>(x: Double,y: Double): org.apache.spark.sql.test.ExamplePoint in
>> class ExamplePoint, tree.tpe=(x: Double, y:
>> Double)org.apache.spark.sql.test.ExamplePoint
>>                 // 2 arguments
>>                 Apply( // def getDouble(x$1: Int): Double in trait
>> SpecializedGetters, tree.tpe=Double
>>                   "x2"."getDouble" // def getDouble(x$1: Int): Double in
>> trait SpecializedGetters, tree.tpe=(x$1: Int)Double
>>                   0
>>                 )
>>                 Apply( // def getDouble(x$1: Int): Double in trait
>> SpecializedGetters, tree.tpe=Double
>>                   "x2"."getDouble" // def getDouble(x$1: Int): Double in
>> trait SpecializedGetters, tree.tpe=(x$1: Int)Double
>>                   1
>>                 )
>>               )
>>             )
>>           )
>>           Apply( // case def case6():
>> org.apache.spark.sql.test.ExamplePoint,
>> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>             "case6" // case def case6():
>> org.apache.spark.sql.test.ExamplePoint,
>> tree.tpe=()org.apache.spark.sql.test.ExamplePoint
>>             Nil
>>           )
>>         )
>>       )
>>       LabelDef( // case def case6():
>> org.apache.spark.sql.test.ExamplePoint,
>> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>         ()
>>         Apply( // case def matchEnd4(x:
>> org.apache.spark.sql.test.ExamplePoint):
>> org.apache.spark.sql.test.ExamplePoint,
>> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>           "matchEnd4" // case def matchEnd4(x:
>> org.apache.spark.sql.test.ExamplePoint):
>> org.apache.spark.sql.test.ExamplePoint, tree.tpe=(x:
>> org.apache.spark.sql.test.ExamplePoint)org.apache.spark.sql.test.ExamplePoint
>>           Throw( // tree.tpe=Nothing
>>             Apply( // def <init>(obj: Object): MatchError in class
>> MatchError, tree.tpe=MatchError
>>               new MatchError."<init>" // def <init>(obj: Object):
>> MatchError in class MatchError, tree.tpe=(obj: Object)MatchError
>>               "x1" // case val x1: Object, tree.tpe=Object
>>             )
>>           )
>>         )
>>       )
>>       LabelDef( // case def matchEnd4(x:
>> org.apache.spark.sql.test.ExamplePoint):
>> org.apache.spark.sql.test.ExamplePoint,
>> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>         "x" // x: org.apache.spark.sql.test.ExamplePoint,
>> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>         "x" // x: org.apache.spark.sql.test.ExamplePoint,
>> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>       )
>>     )
>>   )
>>   DefDef( // override def userClass(): Class in class ExamplePointUDT
>>     <method> override
>>     "userClass"
>>     []
>>     List(Nil)
>>     <tpt> // tree.tpe=Class
>>     classOf[org.apache.spark.sql.test.ExamplePoint]
>>   )
>>   DefDef( // override private[package spark] def asNullable():
>> org.apache.spark.sql.test.ExamplePointUDT in class ExamplePointUDT
>>     <method> override <triedcooking>
>>     "asNullable"
>>     []
>>     List(Nil)
>>     <tpt> // tree.tpe=org.apache.spark.sql.test.ExamplePointUDT
>>     This(<empty>)private[package sql] class ExamplePointUDT extends
>> UserDefinedType in package test,
>> tree.tpe=org.apache.spark.sql.test.ExamplePointUDT
>>   )
>>   DefDef( // override def asNullable():
>> org.apache.spark.sql.types.DataType in class ExamplePointUDT
>>     <method> override <bridge>
>>     "asNullable"
>>     []
>>     List(Nil)
>>     <tpt> // tree.tpe=org.apache.spark.sql.types.DataType
>>     Apply( // override private[package spark] def asNullable():
>> org.apache.spark.sql.test.ExamplePointUDT in class ExamplePointUDT,
>> tree.tpe=org.apache.spark.sql.test.ExamplePointUDT
>>       ExamplePointUDT.this."asNullable" // override private[package
>> spark] def asNullable(): org.apache.spark.sql.test.ExamplePointUDT in class
>> ExamplePointUDT, tree.tpe=()org.apache.spark.sql.test.ExamplePointUDT
>>       Nil
>>     )
>>   )
>>   DefDef( // override def asNullable():
>> org.apache.spark.sql.types.UserDefinedType in class ExamplePointUDT
>>     <method> override <bridge>
>>     "asNullable"
>>     []
>>     List(Nil)
>>     <tpt> // tree.tpe=org.apache.spark.sql.types.UserDefinedType
>>     Apply( // override private[package spark] def asNullable():
>> org.apache.spark.sql.test.ExamplePointUDT in class ExamplePointUDT,
>> tree.tpe=org.apache.spark.sql.test.ExamplePointUDT
>>       ExamplePointUDT.this."asNullable" // override private[package
>> spark] def asNullable(): org.apache.spark.sql.test.ExamplePointUDT in class
>> ExamplePointUDT, tree.tpe=()org.apache.spark.sql.test.ExamplePointUDT
>>       Nil
>>     )
>>   )
>>   DefDef( // override def deserialize(datum: Object): Object in class
>> ExamplePointUDT
>>     <method> override <bridge>
>>     "deserialize"
>>     []
>>     // 1 parameter list
>>     ValDef( // datum: Object
>>       <param> <triedcooking>
>>       "datum"
>>       <tpt> // tree.tpe=Object
>>       <empty>
>>     )
>>     <tpt> // tree.tpe=Object
>>     Apply( // override def deserialize(datum: Object):
>> org.apache.spark.sql.test.ExamplePoint in class ExamplePointUDT,
>> tree.tpe=org.apache.spark.sql.test.ExamplePoint
>>       ExamplePointUDT.this."deserialize" // override def
>> deserialize(datum: Object): org.apache.spark.sql.test.ExamplePoint in class
>> ExamplePointUDT, tree.tpe=(datum:
>> Object)org.apache.spark.sql.test.ExamplePoint
>>       "datum" // datum: Object, tree.tpe=Object
>>     )
>>   )
>>   DefDef( // override def serialize(obj: Object): Object in class
>> ExamplePointUDT
>>     <method> override <bridge>
>>     "serialize"
>>     []
>>     // 1 parameter list
>>     ValDef( // obj: Object
>>       <param> <triedcooking>
>>       "obj"
>>       <tpt> // tree.tpe=Object
>>       <empty>
>>     )
>>     <tpt> // tree.tpe=Object
>>     Apply( // override def serialize(obj: Object):
>> org.apache.spark.sql.catalyst.util.GenericArrayData in class
>> ExamplePointUDT,
>> tree.tpe=org.apache.spark.sql.catalyst.util.GenericArrayData
>>       ExamplePointUDT.this."serialize" // override def serialize(obj:
>> Object): org.apache.spark.sql.catalyst.util.GenericArrayData in class
>> ExamplePointUDT, tree.tpe=(obj:
>> Object)org.apache.spark.sql.catalyst.util.GenericArrayData
>>       "obj" // obj: Object, tree.tpe=Object
>>     )
>>   )
>>   DefDef( // def <init>(): org.apache.spark.sql.test.ExamplePointUDT in
>> class ExamplePointUDT
>>     <method>
>>     "<init>"
>>     []
>>     List(Nil)
>>     <tpt> // tree.tpe=org.apache.spark.sql.test.ExamplePointUDT
>>     Block( // tree.tpe=Unit
>>       Apply( // def <init>(): org.apache.spark.sql.types.UserDefinedType
>> in class UserDefinedType,
>> tree.tpe=org.apache.spark.sql.types.UserDefinedType
>>         ExamplePointUDT.super."<init>" // def <init>():
>> org.apache.spark.sql.types.UserDefinedType in class UserDefinedType,
>> tree.tpe=()org.apache.spark.sql.types.UserDefinedType
>>         Nil
>>       )
>>       ()
>>     )
>>   )
>> )
>> == Expanded type of tree ==
>> *ConstantType(*
>> *  value = Constant(org.apache.spark.sql.test.ExamplePoint)*
>> *)*
>> *uncaught exception during compilation: java.lang.AssertionError*
>>
>> *Error:scala: Error: assertion failed: List(object package$DebugNode,
>> object package$DebugNode)*
>> *java.lang.AssertionError: assertion failed: List(object
>> package$DebugNode, object package$DebugNode)*
>> at scala.reflect.internal.Symbols$Symbol.suchThat(Symbols.scala:1678)
>> at
>> scala.reflect.internal.Symbols$ClassSymbol.companionModule0(Symbols.scala:2988)
>> at
>> scala.reflect.internal.Symbols$ClassSymbol.companionModule(Symbols.scala:2991)
>> at
>> scala.tools.nsc.backend.jvm.GenASM$JPlainBuilder.genClass(GenASM.scala:1371)
>> at scala.tools.nsc.backend.jvm.GenASM$AsmPhase.run(GenASM.scala:120)
>> at scala.tools.nsc.Global$Run.compileUnitsInternal(Global.scala:1583)
>> at scala.tools.nsc.Global$Run.compileUnits(Global.scala:1557)
>> at scala.tools.nsc.Global$Run.compileSources(Global.scala:1553)
>> at scala.tools.nsc.Global$Run.compile(Global.scala:1662)
>> at xsbt.CachedCompiler0.run(CompilerInterface.scala:126)
>> at xsbt.CachedCompiler0.run(CompilerInterface.scala:102)
>> at xsbt.CompilerInterface.run(CompilerInterface.scala:27)
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>> at java.lang.reflect.Method.invoke(Method.java:606)
>> at sbt.compiler.AnalyzingCompiler.call(AnalyzingCompiler.scala:102)
>> at sbt.compiler.AnalyzingCompiler.compile(AnalyzingCompiler.scala:48)
>> at sbt.compiler.AnalyzingCompiler.compile(AnalyzingCompiler.scala:41)
>> at
>> sbt.compiler.AggressiveCompile$$anonfun$6$$anonfun$compileScala$1$1$$anonfun$apply$3$$anonfun$apply$1.apply$mcV$sp(AggressiveCompile.scala:106)
>> at
>> sbt.compiler.AggressiveCompile$$anonfun$6$$anonfun$compileScala$1$1$$anonfun$apply$3$$anonfun$apply$1.apply(AggressiveCompile.scala:106)
>> at
>> sbt.compiler.AggressiveCompile$$anonfun$6$$anonfun$compileScala$1$1$$anonfun$apply$3$$anonfun$apply$1.apply(AggressiveCompile.scala:106)
>> at
>> sbt.compiler.AggressiveCompile.sbt$compiler$AggressiveCompile$$timed(AggressiveCompile.scala:179)
>> at
>> sbt.compiler.AggressiveCompile$$anonfun$6$$anonfun$compileScala$1$1$$anonfun$apply$3.apply(AggressiveCompile.scala:105)
>> at
>> sbt.compiler.AggressiveCompile$$anonfun$6$$anonfun$compileScala$1$1$$anonfun$apply$3.apply(AggressiveCompile.scala:102)
>> at scala.Option.foreach(Option.scala:245)
>> at
>> sbt.compiler.AggressiveCompile$$anonfun$6$$anonfun$compileScala$1$1.apply(AggressiveCompile.scala:102)
>> at
>> sbt.compiler.AggressiveCompile$$anonfun$6$$anonfun$compileScala$1$1.apply(AggressiveCompile.scala:102)
>> at scala.Option.foreach(Option.scala:245)
>> at
>> sbt.compiler.AggressiveCompile$$anonfun$6.compileScala$1(AggressiveCompile.scala:102)
>> at
>> sbt.compiler.AggressiveCompile$$anonfun$6.apply(AggressiveCompile.scala:151)
>> at
>> sbt.compiler.AggressiveCompile$$anonfun$6.apply(AggressiveCompile.scala:89)
>> at sbt.inc.IncrementalCompile$$anonfun$doCompile$1.apply(Compile.scala:40)
>> at sbt.inc.IncrementalCompile$$anonfun$doCompile$1.apply(Compile.scala:38)
>> at sbt.inc.IncrementalCommon.cycle(Incremental.scala:103)
>> at sbt.inc.Incremental$$anonfun$1.apply(Incremental.scala:39)
>> at sbt.inc.Incremental$$anonfun$1.apply(Incremental.scala:38)
>> at sbt.inc.Incremental$.manageClassfiles(Incremental.scala:69)
>> at sbt.inc.Incremental$.compile(Incremental.scala:38)
>> at sbt.inc.IncrementalCompile$.apply(Compile.scala:28)
>> at sbt.compiler.AggressiveCompile.compile2(AggressiveCompile.scala:170)
>> at sbt.compiler.AggressiveCompile.compile1(AggressiveCompile.scala:73)
>> at
>> org.jetbrains.jps.incremental.scala.local.SbtCompiler.compile(SbtCompiler.scala:66)
>> at
>> org.jetbrains.jps.incremental.scala.local.LocalServer.compile(LocalServer.scala:26)
>> at org.jetbrains.jps.incremental.scala.remote.Main$.make(Main.scala:62)
>> at
>> org.jetbrains.jps.incremental.scala.remote.Main$.nailMain(Main.scala:20)
>> at org.jetbrains.jps.incremental.scala.remote.Main.nailMain(Main.scala)
>> at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>> at java.lang.reflect.Method.invoke(Method.java:606)
>> at com.martiansoftware.nailgun.NGSession.run(NGSession.java:319)
>>
>> I just highlighted some error message that I think important as *bold
>> and red.*
>>
>> This really bothered me for several days, I don't know how to get
>> through. Any suggestions? Thanks.
>>
>
>


-- 
Best Regards

Jeff Zhang

Reply via email to