Zinc, I believe, is something you can install and run to speed up your Maven builds. It's not required.
I get a bunch of warnings when compiling with Maven, too. Dunno if they are expected or not, but things work fine from there on. Many people do indeed use sbt. I don't know where we have documentation on how to use sbt (we recently removed it from the README), but sbt/sbt clean followed by sbt/sbt assembly should work fine. Maven is indeed the "proper" way to build Spark, but building with sbt is supported too and most Spark devs I believe use it because it's faster than Maven. Nick On Tue, Nov 4, 2014 at 8:03 PM, Alessandro Baretta <alexbare...@gmail.com> wrote: > Nicholas, > > Yes, I saw them, but they refer to maven, and I'm under the impression > that sbt is the preferred way of building spark. Is indeed maven the "right > way"? Anyway, as per your advice I ctrl-d'ed my sbt shell and have ran `mvn > -DskipTests clean package`, which completed successfully. So, indeed, in > trying to use sbt I was on a wild goose chase. > > Here's a couple of glitches I'm seeing. First of all many warnings such as > the following: > > [WARNING] assert(windowedStream2.generatedRDDs.contains(Time(10000))) > [WARNING] ^ > [WARNING] > /home/alex/git/spark/streaming/src/test/scala/org/apache/spark/streaming/BasicOperationsSuite.scala:454: > inferred existential type > scala.collection.mutable.HashMap[org.apache.spark.streaming.Time,org.apache.spark.rdd.RDD[_$2]] > forSome { type _$2 }, which cannot be expressed by wildcards, should be > enabled > by making the implicit value scala.language.existentials visible. > > [WARNING] > /home/alex/git/spark/sql/hive/src/main/scala/org/apache/spark/sql/hive/parquet/FakeParquetSerDe.scala:34: > @deprecated now takes two arguments; see the scaladoc. > [WARNING] @deprecated("No code should depend on FakeParquetHiveSerDe as it > is only intended as a " + > [WARNING] ^ > > [WARNING] > /home/alex/git/spark/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveMetastoreCatalog.scala:435: > trait Deserializer in package serde2 is deprecated: see corresponding > Javadoc for more information. > [WARNING] > Utils.getContextOrSparkClassLoader).asInstanceOf[Class[Deserializer]], > [WARNING] ^ > > [WARNING] > /home/alex/git/spark/examples/src/main/scala/org/apache/spark/examples/mllib/StreamingKMeans.scala:22: > imported `StreamingKMeans' is permanently hidden by definition of object > StreamingKMeans in package mllib > [WARNING] import org.apache.spark.mllib.clustering.StreamingKMeans > Are they expected? > > Also, mvn complains about not having zinc. Is this a problem? > > [WARNING] Zinc server is not available at port 3030 - reverting to normal > incremental compile > > Alex > > On Tue, Nov 4, 2014 at 3:11 PM, Nicholas Chammas < > nicholas.cham...@gmail.com> wrote: > >> FWIW, the "official" build instructions are here: >> https://github.com/apache/spark#building-spark >> >> On Tue, Nov 4, 2014 at 5:11 PM, Ted Yu <yuzhih...@gmail.com> wrote: >> >>> I built based on this commit today and the build was successful. >>> >>> What command did you use ? >>> >>> Cheers >>> >>> On Tue, Nov 4, 2014 at 2:08 PM, Alessandro Baretta < >>> alexbare...@gmail.com> >>> wrote: >>> >>> > Fellow Sparkers, >>> > >>> > I am new here and still trying to learn to crawl. Please, bear with me. >>> > >>> > I just pulled f90ad5d from https://github.com/apache/spark.git and am >>> > running the compile command in the sbt shell. This is the error I'm >>> seeing: >>> > >>> > [error] >>> > >>> > >>> /home/alex/git/spark/mllib/src/main/scala/org/apache/spark/mllib/linalg/Vectors.scala:32: >>> > object sql is not a member of package org.apache.spark >>> > [error] import org.apache.spark.sql.catalyst.types._ >>> > [error] ^ >>> > >>> > Am I doing something obscenely stupid is the build genuinely broken? >>> > >>> > Alex >>> > >>> >> >> >