Why do you need to skip java tests? I build the distro just fine with Java 8. On Dec 27, 2014 4:21 AM, "Ted Yu" <yuzhih...@gmail.com> wrote:
> In case jdk 1.7 or higher is used to build, --skip-java-test needs to be > specifed. > > FYI > > On Thu, Dec 25, 2014 at 5:03 PM, guxiaobo1982 <guxiaobo1...@qq.com> wrote: > >> The following command works >> >> ./make-distribution.sh --tgz -Pyarn -Dyarn.version=2.6.0 -Phadoop-2.4 >> -Dhadoop.version=2.6.0 -Phive -DskipTests >> >> ------------------ Original ------------------ >> *From: * "guxiaobo1982";<guxiaobo1...@qq.com>; >> *Send time:* Thursday, Dec 25, 2014 3:58 PM >> *To:* ""<guxiaobo1...@qq.com>; "Ted Yu"<yuzhih...@gmail.com>; >> *Cc:* "user@spark.apache.org"<user@spark.apache.org>; >> *Subject: * Re: How to build Spark against the latest >> >> >> What options should I use when running the make-distribution.sh script, >> >> I tried ./make-distribution.sh --hadoop.version 2.6.0 --with-yarn >> -with-hive --with-tachyon --tgz >> with nothing came out. >> >> Regards >> >> ------------------ Original ------------------ >> *From: * "guxiaobo1982";<guxiaobo1...@qq.com>; >> *Send time:* Wednesday, Dec 24, 2014 6:52 PM >> *To:* "Ted Yu"<yuzhih...@gmail.com>; >> *Cc:* "user@spark.apache.org"<user@spark.apache.org>; >> *Subject: * Re: How to build Spark against the latest >> >> Hi Ted, >> The reference command works, but where I can get the deployable >> binaries? >> >> Xiaobo Gu >> >> >> >> >> ------------------ Original ------------------ >> *From: * "Ted Yu";<yuzhih...@gmail.com>; >> *Send time:* Wednesday, Dec 24, 2014 12:09 PM >> *To:* ""<guxiaobo1...@qq.com>; >> *Cc:* "user@spark.apache.org"<user@spark.apache.org>; >> *Subject: * Re: How to build Spark against the latest >> >> See http://search-hadoop.com/m/JW1q5Cew0j >> >> On Tue, Dec 23, 2014 at 8:00 PM, guxiaobo1982 <guxiaobo1...@qq.com> >> wrote: >> >>> Hi, >>> The official pom.xml file only have profile for hadoop version 2.4 as >>> the latest version, but I installed hadoop version 2.6.0 with ambari, how >>> can I build spark against it, just using mvn -Dhadoop.version=2.6.0, or >>> how to make a coresponding profile for it? >>> >>> Regards, >>> >>> Xiaobo >>> >> >> >