try `sbt/sbt clean` first? -Xiangrui
On Wed, Jul 23, 2014 at 11:23 AM, m3.sharma <sharm...@umn.edu> wrote: > I am trying to build spark after cloning from github repo: > > I am executing: > ./sbt/sbt -Dhadoop.version=2.4.0 -Pyarn assembly > > I am getting following error: > [warn] ^ > [error] > [error] while compiling: > /home/m3.sharma/installSrc/spark/spark/sql/core/src/main/scala/org/apache/spark/sql/test/TestSQLContext.scala > [error] during phase: jvm > [error] library version: version 2.10.4 > [error] compiler version: version 2.10.4 > [error] reconstructed args: -classpath /home/ ......... > .... > [error] > [error] last tree to typer: > Literal(Constant(org.apache.spark.sql.catalyst.types.PrimitiveType)) > [error] symbol: null > [error] symbol definition: null > [error] tpe: > Class(classOf[org.apache.spark.sql.catalyst.types.PrimitiveType]) > [error] symbol owners: > [error] context owners: object TestSQLContext -> package test > [error] > [error] == Enclosing template or block == > [error] > [error] Template( // val <local TestSQLContext>: <notype> in object > TestSQLContext, tree.tpe=org.apache.spark.sql.test.TestSQLContext.type > [error] "org.apache.spark.sql.SQLContext" // parents > [error] ValDef( > [error] private > [error] "_" > [error] <tpt> > [error] <empty> > [error] ) > [error] // 2 statements > [error] DefDef( // private def readResolve(): Object in object > TestSQLContext > [error] <method> private <synthetic> > [error] "readResolve" > [error] [] > [error] List(Nil) > [error] <tpt> // tree.tpe=Object > [error] test.this."TestSQLContext" // object TestSQLContext in package > test, tree.tpe=org.apache.spark.sql.test.TestSQLContext.type > [error] ) > [error] DefDef( // def <init>(): > org.apache.spark.sql.test.TestSQLContext.type in object TestSQLContext > [error] <method> > [error] "<init>" > [error] [] > [error] List(Nil) > [error] <tpt> // tree.tpe=org.apache.spark.sql.test.TestSQLContext.type > [error] Block( // tree.tpe=Unit > [error] Apply( // def <init>(sparkContext: > org.apache.spark.SparkContext): org.apache.spark.sql.SQLContext in class > SQLContext, tree.tpe=org.apach > e.spark.sql.SQLContext > [error] TestSQLContext.super."<init>" // def <init>(sparkContext: > org.apache.spark.SparkContext): org.apache.spark.sql.SQLContext in class > SQLCo > ntext, tree.tpe=(sparkContext: > org.apache.spark.SparkContext)org.apache.spark.sql.SQLContext > [error] Apply( // def <init>(master: String,appName: String,conf: > org.apache.spark.SparkConf): org.apache.spark.SparkContext in class > SparkConte > xt, tree.tpe=org.apache.spark.SparkContext > [error] new org.apache.spark.SparkContext."<init>" // def > <init>(master: String,appName: String,conf: org.apache.spark.SparkConf): > org.apache. > spark.SparkContext in class SparkContext, tree.tpe=(master: String, appName: > String, conf: org.apache.spark.SparkConf)org.apache.spark.SparkContext > [error] // 3 arguments > [error] "local" > [error] "TestSQLContext" > [error] Apply( // def <init>(): org.apache.spark.SparkConf in > class SparkConf, tree.tpe=org.apache.spark.SparkConf > [error] new org.apache.spark.SparkConf."<init>" // def <init>(): > org.apache.spark.SparkConf in class SparkConf, tree.tpe=()org.apache.spark. > SparkConf > [error] Nil > [error] ) > [error] ) > [error] ) > [error] () > [error] ) > [error] ) > [error] ) > [error] > [error] == Expanded type of tree == > [error] > [error] ConstantType( > [error] value = > Constant(org.apache.spark.sql.catalyst.types.PrimitiveType) > [error] ) > [error] > [error] uncaught exception during compilation: java.lang.AssertionError > java.lang.AssertionError: assertion failed: List(object package$DebugNode, > object package$DebugNode) > at scala.reflect.internal.Symbols$Symbol.suchThat(Symbols.scala:1678) > at > scala.reflect.internal.Symbols$ClassSymbol.companionModule0(Symbols.scala:2988) > at > scala.reflect.internal.Symbols$ClassSymbol.companionModule(Symbols.scala:2991) > at > scala.tools.nsc.backend.jvm.GenASM$JPlainBuilder.genClass(GenASM.scala:1371) > at scala.tools.nsc.backend.jvm.GenASM$AsmPhase.run(GenASM.scala:120) > at > scala.tools.nsc.Global$Run.compileUnitsInternal(Global.scala:1583) > at scala.tools.nsc.Global$Run.compileUnits(Global.scala:1557) > at scala.tools.nsc.Global$Run.compileSources(Global.scala:1553) > at scala.tools.nsc.Global$Run.compile(Global.scala:1662) > at xsbt.CachedCompiler0.run(CompilerInterface.scala:123) > at xsbt.CachedCompiler0.run(CompilerInterface.scala:99) > at xsbt.CompilerInterface.run(CompilerInterface.scala:27) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at sbt.compiler.AnalyzingCompiler.call(AnalyzingCompiler.scala:102) > at > sbt.compiler.AnalyzingCompiler.compile(AnalyzingCompiler.scala:48) > at > sbt.compiler.AnalyzingCompiler.compile(AnalyzingCompiler.scala:41) > at > sbt.compiler.AggressiveCompile$$anonfun$3$$anonfun$compileScala$1$1.apply$mcV$sp(AggressiveCompile.scala:99) > at > sbt.compiler.AggressiveCompile$$anonfun$3$$anonfun$compileScala$1$1.apply(AggressiveCompile.scala:99) > at > sbt.compiler.AggressiveCompile$$anonfun$3$$anonfun$compileScala$1$1.apply(AggressiveCompile.scala:99) > at > sbt.compiler.AggressiveCompile.sbt$compiler$AggressiveCompile$$timed(AggressiveCompile.scala:166) > at > sbt.compiler.AggressiveCompile$$anonfun$3.compileScala$1(AggressiveCompile.scala:98) > at > sbt.compiler.AggressiveCompile$$anonfun$3.apply(AggressiveCompile.scala:143) > at > sbt.compiler.AggressiveCompile$$anonfun$3.apply(AggressiveCompile.scala:87) > at > sbt.inc.IncrementalCompile$$anonfun$doCompile$1.apply(Compile.scala:39) > at > sbt.inc.IncrementalCompile$$anonfun$doCompile$1.apply(Compile.scala:37) > at sbt.inc.IncrementalCommon.cycle(Incremental.scala:99) > at sbt.inc.Incremental$$anonfun$1.apply(Incremental.scala:38) > at sbt.inc.Incremental$$anonfun$1.apply(Incremental.scala:37) > at sbt.inc.Incremental$.manageClassfiles(Incremental.scala:65) > at sbt.inc.Incremental$.compile(Incremental.scala:37) > at sbt.inc.IncrementalCompile$.apply(Compile.scala:27) > at > sbt.compiler.AggressiveCompile.compile2(AggressiveCompile.scala:157) > at > sbt.compiler.AggressiveCompile.compile1(AggressiveCompile.scala:71) > at sbt.compiler.AggressiveCompile.apply(AggressiveCompile.scala:46) > at sbt.Compiler$.apply(Compiler.scala:75) > at sbt.Compiler$.apply(Compiler.scala:66) > at sbt.Defaults$.sbt$Defaults$$compileTaskImpl(Defaults.scala:770) > at sbt.Defaults$$anonfun$compileTask$1.apply(Defaults.scala:762) > at sbt.Defaults$$anonfun$compileTask$1.apply(Defaults.scala:762) > at scala.Function1$$anonfun$compose$1.apply(Function1.scala:47) > at > sbt.$tilde$greater$$anonfun$$u2219$1.apply(TypeFunctions.scala:42) > at sbt.std.Transform$$anon$4.work(System.scala:64) > at > sbt.Execute$$anonfun$submit$1$$anonfun$apply$1.apply(Execute.scala:237) > at > sbt.Execute$$anonfun$submit$1$$anonfun$apply$1.apply(Execute.scala:237) > at sbt.ErrorHandling$.wideConvert(ErrorHandling.scala:18) > at sbt.Execute.work(Execute.scala:244) > at sbt.Execute$$anonfun$submit$1.apply(Execute.scala:237) > at sbt.Execute$$anonfun$submit$1.apply(Execute.scala:237) > at > sbt.ConcurrentRestrictions$$anon$4$$anonfun$1.apply(ConcurrentRestrictions.scala:160) > at sbt.CompletionService$$anon$2.call(CompletionService.scala:30) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:744) > [error] (sql/compile:compile) java.lang.AssertionError: assertion failed: > List(object package$DebugNode, object package$DebugNode) > [error] Total time: 126 s, completed Jul 23, 2014 11:19:27 AM > > I dont want spark sql, I can do without it. > > > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/spark-github-source-build-error-tp10532.html > Sent from the Apache Spark User List mailing list archive at Nabble.com.