hadoop-2.6 is supported (look for "profile" XML in the pom.xml file).

For Hive, add "-Phive -Phive-thriftserver"  (See
http://spark.apache.org/docs/latest/sql-programming-guide.html#hive-tables)
for more details.

dean

Dean Wampler, Ph.D.
Author: Programming Scala, 2nd Edition
<http://shop.oreilly.com/product/0636920033073.do> (O'Reilly)
Typesafe <http://typesafe.com>
@deanwampler <http://twitter.com/deanwampler>
http://polyglotprogramming.com

On Mon, Jul 20, 2015 at 2:55 PM, Michael Segel <msegel_had...@hotmail.com>
wrote:

> Sorry,
>
> Should have sent this to user…
>
> However… it looks like the docs page may need some editing?
>
> Thx
>
> -Mike
>
>
> Begin forwarded message:
>
> *From: *Michael Segel <msegel_had...@hotmail.com>
> *Subject: **Silly question about building Spark 1.4.1*
> *Date: *July 20, 2015 at 12:26:40 PM MST
> *To: *d...@spark.apache.org
>
> Hi,
>
> I’m looking at the online docs for building spark 1.4.1 …
>
> http://spark.apache.org/docs/latest/building-spark.html
>
> I was interested in building spark for Scala 2.11 (latest scala) and also
> for Hive and JDBC support.
>
> The docs say:
> “
> To produce a Spark package compiled with Scala 2.11, use the -Dscala-2.11
> property:
>
> dev/change-version-to-2.11.sh
> mvn -Pyarn -Phadoop-2.4 -Dscala-2.11 -DskipTests clean package
>
> “
>
> So…
>
> Is there a reason I shouldn’t build against hadoop-2.6 ?
>
>
> If I want to add the Thirft and Hive support, is it possible?
>
> Looking at the Scala build, it looks like hive support is being built?
>
> (Looking at the stdout messages…)
>
> Should the docs be updated? Am I missing something?
>
> (Dean W. can confirm, I am completely brain dead. ;-)
>
>
> Thx
>
>
> -Mike
>
> PS. Yes I can probably download a prebuilt image, but I’m a glutton for 
> punishment. ;-)
>
>
>
>

Reply via email to