Hi Nathan,

try 'sbt clean' and then re-run make-distribution.sh.

To clean really every trace of previous builds, try also removing ~/.m2,
~/.ivy2, ~/.sbt. That's a little time expensive, cause all external
artifacts will be downloaded again, but does the trick.

Kind regards,
Eugene


On Thu, Nov 28, 2013 at 7:18 PM, Nathan Kronenfeld <
[email protected]> wrote:

> Hi, folks.
>
> I'm trying to build the a spark distribution with the latest code.
>
> I started out this morning with:
>
> ./make-distribution.sh
>
>
> and that worked fine. But then I realized I'd forgotten to set the hadoop
> version I needed, so I redid it with
>
> ./make-distribution.sh --hadoop 2.0.0-cdh4.4.0
>
>
> That failed with a whole bunch of error messages (43 to be exact) in
> streaming on the lines of:
>
> ...streaming/src/main/scala/org/apache/spark/streaming/api/java/JavaPairDStream.scala:51:
> type mismatch
> found: org.apache.spark.streaming..DStream[(K, V)]
> expected: org.apache.spark.streaming.api.java.JavaPairDStream[K, V]
> Note: implicit method fromPairDStream is not applicable here because it
> comes after the application point and it lacks an explicit return type.
> dstream.filter(x => f(x).booleanValue())
>
>
> (42 more like that in different places).  So I went back and tried
>
> ./make-distribution.sh
>
>
> again - now it failed with the same errors, though it just worked a second
> ago.  Clean up the dist directory - same thing  Log out and in to reset my
> environment - same thing.
>
> So though it built fine once, now it refuses to build again.
>
> Does anyone have a clue what is going on here?
>
> Any help very much appreciated,
>                 -Nathan
>
>
> --
> Nathan Kronenfeld
> Senior Visualization Developer
> Oculus Info Inc
> 2 Berkeley Street, Suite 600,
> Toronto, Ontario M5A 4J5
> Phone:  +1-416-203-3003 x 238
> Email:  [email protected]
>

Reply via email to