Hi Jacek,
it works fine with mvn: the problem is with sbt.
I suspect a different reactor order in sbt compare to mvn.
Regards
JB
On 11/03/2015 02:44 PM, Jacek Laskowski wrote:
Hi,
Just built the sources using the following command and it worked fine.
➜ spark git:(master) ✗ ./build/mvn -Pyarn -Phadoop-2.6
-Dhadoop.version=2.7.1 -Dscala-2.11 -Phive -Phive-thriftserver
-DskipTests clean install
...
[INFO] ------------------------------------------------------------------------
[INFO] BUILD SUCCESS
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 14:15 min
[INFO] Finished at: 2015-11-03T14:40:40+01:00
[INFO] Final Memory: 438M/1972M
[INFO] ------------------------------------------------------------------------
➜ spark git:(master) ✗ java -version
java version "1.8.0_66"
Java(TM) SE Runtime Environment (build 1.8.0_66-b17)
Java HotSpot(TM) 64-Bit Server VM (build 25.66-b17, mixed mode)
I'm on Mac OS.
Pozdrawiam,
Jacek
--
Jacek Laskowski | http://blog.japila.pl | http://blog.jaceklaskowski.pl
Follow me at https://twitter.com/jaceklaskowski
Upvote at http://stackoverflow.com/users/1305344/jacek-laskowski
On Tue, Nov 3, 2015 at 1:37 PM, Jean-Baptiste Onofré <j...@nanthrax.net> wrote:
Thanks for the update, I used mvn to build but without hive profile.
Let me try with mvn with the same options as you and sbt also.
I keep you posted.
Regards
JB
On 11/03/2015 12:55 PM, Jeff Zhang wrote:
I found it is due to SPARK-11073.
Here's the command I used to build
build/sbt clean compile -Pyarn -Phadoop-2.6 -Phive -Phive-thriftserver
-Psparkr
On Tue, Nov 3, 2015 at 7:52 PM, Jean-Baptiste Onofré <j...@nanthrax.net
<mailto:j...@nanthrax.net>> wrote:
Hi Jeff,
it works for me (with skipping the tests).
Let me try again, just to be sure.
Regards
JB
On 11/03/2015 11:50 AM, Jeff Zhang wrote:
Looks like it's due to guava version conflicts, I see both guava
14.0.1
and 16.0.1 under lib_managed/bundles. Anyone meet this issue too ?
[error]
/Users/jzhang/github/spark_apache/core/src/main/scala/org/apache/spark/SecurityManager.scala:26:
object HashCodes is not a member of package com.google.common.hash
[error] import com.google.common.hash.HashCodes
[error] ^
[info] Resolving org.apache.commons#commons-math;2.2 ...
[error]
/Users/jzhang/github/spark_apache/core/src/main/scala/org/apache/spark/SecurityManager.scala:384:
not found: value HashCodes
[error] val cookie =
HashCodes.fromBytes(secret).toString()
[error] ^
--
Best Regards
Jeff Zhang
--
Jean-Baptiste Onofré
jbono...@apache.org <mailto:jbono...@apache.org>
http://blog.nanthrax.net
Talend - http://www.talend.com
---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
<mailto:dev-unsubscr...@spark.apache.org>
For additional commands, e-mail: dev-h...@spark.apache.org
<mailto:dev-h...@spark.apache.org>
--
Best Regards
Jeff Zhang
--
Jean-Baptiste Onofré
jbono...@apache.org
http://blog.nanthrax.net
Talend - http://www.talend.com
---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
For additional commands, e-mail: dev-h...@spark.apache.org
---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
For additional commands, e-mail: dev-h...@spark.apache.org
--
Jean-Baptiste Onofré
jbono...@apache.org
http://blog.nanthrax.net
Talend - http://www.talend.com
---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
For additional commands, e-mail: dev-h...@spark.apache.org