Zinc, I believe, is something you can install and run to speed up your
Maven builds. It's not required.

I get a bunch of warnings when compiling with Maven, too. Dunno if they are
expected or not, but things work fine from there on.

Many people do indeed use sbt. I don't know where we have documentation on
how to use sbt (we recently removed it from the README), but sbt/sbt clean
followed by sbt/sbt assembly should work fine.

Maven is indeed the "proper" way to build Spark, but building with sbt is
supported too and most Spark devs I believe use it because it's faster than
Maven.

Nick

On Tue, Nov 4, 2014 at 8:03 PM, Alessandro Baretta <alexbare...@gmail.com>
wrote:

> Nicholas,
>
> Yes, I saw them, but they refer to maven, and I'm under the impression
> that sbt is the preferred way of building spark. Is indeed maven the "right
> way"? Anyway, as per your advice I ctrl-d'ed my sbt shell and have ran `mvn
> -DskipTests clean package`, which completed successfully. So, indeed, in
> trying to use sbt I was on a wild goose chase.
>
> Here's a couple of glitches I'm seeing. First of all many warnings such as
> the following:
>
> [WARNING]     assert(windowedStream2.generatedRDDs.contains(Time(10000)))
> [WARNING]                            ^
> [WARNING]
> /home/alex/git/spark/streaming/src/test/scala/org/apache/spark/streaming/BasicOperationsSuite.scala:454:
> inferred existential type
> scala.collection.mutable.HashMap[org.apache.spark.streaming.Time,org.apache.spark.rdd.RDD[_$2]]
> forSome { type _$2 }, which cannot be expressed by wildcards,  should be
> enabled
> by making the implicit value scala.language.existentials visible.
>
> [WARNING]
> /home/alex/git/spark/sql/hive/src/main/scala/org/apache/spark/sql/hive/parquet/FakeParquetSerDe.scala:34:
> @deprecated now takes two arguments; see the scaladoc.
> [WARNING] @deprecated("No code should depend on FakeParquetHiveSerDe as it
> is only intended as a " +
> [WARNING]  ^
>
> [WARNING]
> /home/alex/git/spark/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveMetastoreCatalog.scala:435:
> trait Deserializer in package serde2 is deprecated: see corresponding
> Javadoc for more information.
> [WARNING]
> Utils.getContextOrSparkClassLoader).asInstanceOf[Class[Deserializer]],
> [WARNING]                                                        ^
>
> [WARNING]
> /home/alex/git/spark/examples/src/main/scala/org/apache/spark/examples/mllib/StreamingKMeans.scala:22:
> imported `StreamingKMeans' is permanently hidden by definition of object
> StreamingKMeans in package mllib
> [WARNING] import org.apache.spark.mllib.clustering.StreamingKMeans
> Are they expected?
>
> Also, mvn complains about not having zinc. Is this a problem?
>
> [WARNING] Zinc server is not available at port 3030 - reverting to normal
> incremental compile
>
> Alex
>
> On Tue, Nov 4, 2014 at 3:11 PM, Nicholas Chammas <
> nicholas.cham...@gmail.com> wrote:
>
>> FWIW, the "official" build instructions are here:
>> https://github.com/apache/spark#building-spark
>>
>> On Tue, Nov 4, 2014 at 5:11 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>
>>> I built based on this commit today and the build was successful.
>>>
>>> What command did you use ?
>>>
>>> Cheers
>>>
>>> On Tue, Nov 4, 2014 at 2:08 PM, Alessandro Baretta <
>>> alexbare...@gmail.com>
>>> wrote:
>>>
>>> > Fellow Sparkers,
>>> >
>>> > I am new here and still trying to learn to crawl. Please, bear with me.
>>> >
>>> > I just pulled f90ad5d from https://github.com/apache/spark.git and am
>>> > running the compile command in the sbt shell. This is the error I'm
>>> seeing:
>>> >
>>> > [error]
>>> >
>>> >
>>> /home/alex/git/spark/mllib/src/main/scala/org/apache/spark/mllib/linalg/Vectors.scala:32:
>>> > object sql is not a member of package org.apache.spark
>>> > [error] import org.apache.spark.sql.catalyst.types._
>>> > [error]                         ^
>>> >
>>> > Am I doing something obscenely stupid is the build genuinely broken?
>>> >
>>> > Alex
>>> >
>>>
>>
>>
>

Reply via email to