try `sbt/sbt clean` first? -Xiangrui

On Wed, Jul 23, 2014 at 11:23 AM, m3.sharma <sharm...@umn.edu> wrote:
> I am trying to build spark after cloning from github repo:
>
> I am executing:
> ./sbt/sbt -Dhadoop.version=2.4.0 -Pyarn assembly
>
> I am getting following error:
> [warn]                                 ^
> [error]
> [error]      while compiling:
> /home/m3.sharma/installSrc/spark/spark/sql/core/src/main/scala/org/apache/spark/sql/test/TestSQLContext.scala
> [error]         during phase: jvm
> [error]      library version: version 2.10.4
> [error]     compiler version: version 2.10.4
> [error]   reconstructed args: -classpath /home/ .........
> ....
> [error]
> [error]   last tree to typer:
> Literal(Constant(org.apache.spark.sql.catalyst.types.PrimitiveType))
> [error]               symbol: null
> [error]    symbol definition: null
> [error]                  tpe:
> Class(classOf[org.apache.spark.sql.catalyst.types.PrimitiveType])
> [error]        symbol owners:
> [error]       context owners: object TestSQLContext -> package test
> [error]
> [error] == Enclosing template or block ==
> [error]
> [error] Template( // val <local TestSQLContext>: <notype> in object
> TestSQLContext, tree.tpe=org.apache.spark.sql.test.TestSQLContext.type
> [error]   "org.apache.spark.sql.SQLContext" // parents
> [error]   ValDef(
> [error]     private
> [error]     "_"
> [error]     <tpt>
> [error]     <empty>
> [error]   )
> [error]   // 2 statements
> [error]   DefDef( // private def readResolve(): Object in object
> TestSQLContext
> [error]     <method> private <synthetic>
> [error]     "readResolve"
> [error]     []
> [error]     List(Nil)
> [error]     <tpt> // tree.tpe=Object
> [error]     test.this."TestSQLContext" // object TestSQLContext in package
> test, tree.tpe=org.apache.spark.sql.test.TestSQLContext.type
> [error]   )
> [error]   DefDef( // def <init>():
> org.apache.spark.sql.test.TestSQLContext.type in object TestSQLContext
> [error]     <method>
> [error]     "<init>"
> [error]     []
> [error]     List(Nil)
> [error]     <tpt> // tree.tpe=org.apache.spark.sql.test.TestSQLContext.type
> [error]     Block( // tree.tpe=Unit
> [error]       Apply( // def <init>(sparkContext:
> org.apache.spark.SparkContext): org.apache.spark.sql.SQLContext in class
> SQLContext, tree.tpe=org.apach
> e.spark.sql.SQLContext
> [error]         TestSQLContext.super."<init>" // def <init>(sparkContext:
> org.apache.spark.SparkContext): org.apache.spark.sql.SQLContext in class
> SQLCo
> ntext, tree.tpe=(sparkContext:
> org.apache.spark.SparkContext)org.apache.spark.sql.SQLContext
> [error]         Apply( // def <init>(master: String,appName: String,conf:
> org.apache.spark.SparkConf): org.apache.spark.SparkContext in class
> SparkConte
> xt, tree.tpe=org.apache.spark.SparkContext
> [error]           new org.apache.spark.SparkContext."<init>" // def
> <init>(master: String,appName: String,conf: org.apache.spark.SparkConf):
> org.apache.
> spark.SparkContext in class SparkContext, tree.tpe=(master: String, appName:
> String, conf: org.apache.spark.SparkConf)org.apache.spark.SparkContext
> [error]           // 3 arguments
> [error]           "local"
> [error]           "TestSQLContext"
> [error]           Apply( // def <init>(): org.apache.spark.SparkConf in
> class SparkConf, tree.tpe=org.apache.spark.SparkConf
> [error]             new org.apache.spark.SparkConf."<init>" // def <init>():
> org.apache.spark.SparkConf in class SparkConf, tree.tpe=()org.apache.spark.
> SparkConf
> [error]             Nil
> [error]           )
> [error]         )
> [error]       )
> [error]       ()
> [error]     )
> [error]   )
> [error] )
> [error]
> [error] == Expanded type of tree ==
> [error]
> [error] ConstantType(
> [error]   value =
> Constant(org.apache.spark.sql.catalyst.types.PrimitiveType)
> [error] )
> [error]
> [error] uncaught exception during compilation: java.lang.AssertionError
> java.lang.AssertionError: assertion failed: List(object package$DebugNode,
> object package$DebugNode)
>   at scala.reflect.internal.Symbols$Symbol.suchThat(Symbols.scala:1678)
>         at
> scala.reflect.internal.Symbols$ClassSymbol.companionModule0(Symbols.scala:2988)
>         at
> scala.reflect.internal.Symbols$ClassSymbol.companionModule(Symbols.scala:2991)
>         at
> scala.tools.nsc.backend.jvm.GenASM$JPlainBuilder.genClass(GenASM.scala:1371)
>         at scala.tools.nsc.backend.jvm.GenASM$AsmPhase.run(GenASM.scala:120)
>         at
> scala.tools.nsc.Global$Run.compileUnitsInternal(Global.scala:1583)
>         at scala.tools.nsc.Global$Run.compileUnits(Global.scala:1557)
>         at scala.tools.nsc.Global$Run.compileSources(Global.scala:1553)
>         at scala.tools.nsc.Global$Run.compile(Global.scala:1662)
>         at xsbt.CachedCompiler0.run(CompilerInterface.scala:123)
>         at xsbt.CachedCompiler0.run(CompilerInterface.scala:99)
>         at xsbt.CompilerInterface.run(CompilerInterface.scala:27)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:606)
>         at sbt.compiler.AnalyzingCompiler.call(AnalyzingCompiler.scala:102)
>         at
> sbt.compiler.AnalyzingCompiler.compile(AnalyzingCompiler.scala:48)
>         at
> sbt.compiler.AnalyzingCompiler.compile(AnalyzingCompiler.scala:41)
>         at
> sbt.compiler.AggressiveCompile$$anonfun$3$$anonfun$compileScala$1$1.apply$mcV$sp(AggressiveCompile.scala:99)
>         at
> sbt.compiler.AggressiveCompile$$anonfun$3$$anonfun$compileScala$1$1.apply(AggressiveCompile.scala:99)
>         at
> sbt.compiler.AggressiveCompile$$anonfun$3$$anonfun$compileScala$1$1.apply(AggressiveCompile.scala:99)
>         at
> sbt.compiler.AggressiveCompile.sbt$compiler$AggressiveCompile$$timed(AggressiveCompile.scala:166)
>         at
> sbt.compiler.AggressiveCompile$$anonfun$3.compileScala$1(AggressiveCompile.scala:98)
>         at
> sbt.compiler.AggressiveCompile$$anonfun$3.apply(AggressiveCompile.scala:143)
>         at
> sbt.compiler.AggressiveCompile$$anonfun$3.apply(AggressiveCompile.scala:87)
>         at
> sbt.inc.IncrementalCompile$$anonfun$doCompile$1.apply(Compile.scala:39)
>         at
> sbt.inc.IncrementalCompile$$anonfun$doCompile$1.apply(Compile.scala:37)
>         at sbt.inc.IncrementalCommon.cycle(Incremental.scala:99)
>         at sbt.inc.Incremental$$anonfun$1.apply(Incremental.scala:38)
>         at sbt.inc.Incremental$$anonfun$1.apply(Incremental.scala:37)
>         at sbt.inc.Incremental$.manageClassfiles(Incremental.scala:65)
>         at sbt.inc.Incremental$.compile(Incremental.scala:37)
>         at sbt.inc.IncrementalCompile$.apply(Compile.scala:27)
>         at
> sbt.compiler.AggressiveCompile.compile2(AggressiveCompile.scala:157)
>         at
> sbt.compiler.AggressiveCompile.compile1(AggressiveCompile.scala:71)
>         at sbt.compiler.AggressiveCompile.apply(AggressiveCompile.scala:46)
>         at sbt.Compiler$.apply(Compiler.scala:75)
>         at sbt.Compiler$.apply(Compiler.scala:66)
>         at sbt.Defaults$.sbt$Defaults$$compileTaskImpl(Defaults.scala:770)
>         at sbt.Defaults$$anonfun$compileTask$1.apply(Defaults.scala:762)
>         at sbt.Defaults$$anonfun$compileTask$1.apply(Defaults.scala:762)
>         at scala.Function1$$anonfun$compose$1.apply(Function1.scala:47)
>         at
> sbt.$tilde$greater$$anonfun$$u2219$1.apply(TypeFunctions.scala:42)
>         at sbt.std.Transform$$anon$4.work(System.scala:64)
>         at
> sbt.Execute$$anonfun$submit$1$$anonfun$apply$1.apply(Execute.scala:237)
>         at
> sbt.Execute$$anonfun$submit$1$$anonfun$apply$1.apply(Execute.scala:237)
>         at sbt.ErrorHandling$.wideConvert(ErrorHandling.scala:18)
>         at sbt.Execute.work(Execute.scala:244)
>         at sbt.Execute$$anonfun$submit$1.apply(Execute.scala:237)
>         at sbt.Execute$$anonfun$submit$1.apply(Execute.scala:237)
>         at
> sbt.ConcurrentRestrictions$$anon$4$$anonfun$1.apply(ConcurrentRestrictions.scala:160)
>         at sbt.CompletionService$$anon$2.call(CompletionService.scala:30)
>         at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>         at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)
>         at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>         at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>         at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>         at java.lang.Thread.run(Thread.java:744)
> [error] (sql/compile:compile) java.lang.AssertionError: assertion failed:
> List(object package$DebugNode, object package$DebugNode)
> [error] Total time: 126 s, completed Jul 23, 2014 11:19:27 AM
>
> I dont want spark sql, I can do without it.
>
>
>
>
>
> --
> View this message in context: 
> http://apache-spark-user-list.1001560.n3.nabble.com/spark-github-source-build-error-tp10532.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.

Reply via email to