Why do you need to skip java tests? I build the distro just fine with Java
8.
On Dec 27, 2014 4:21 AM, "Ted Yu" <yuzhih...@gmail.com> wrote:

> In case jdk 1.7 or higher is used to build, --skip-java-test needs to be
> specifed.
>
> FYI
>
> On Thu, Dec 25, 2014 at 5:03 PM, guxiaobo1982 <guxiaobo1...@qq.com> wrote:
>
>> The following command works
>>
>> ./make-distribution.sh --tgz  -Pyarn -Dyarn.version=2.6.0 -Phadoop-2.4
>> -Dhadoop.version=2.6.0 -Phive -DskipTests
>>
>> ------------------ Original ------------------
>> *From: * "guxiaobo1982";<guxiaobo1...@qq.com>;
>> *Send time:* Thursday, Dec 25, 2014 3:58 PM
>> *To:* ""<guxiaobo1...@qq.com>; "Ted Yu"<yuzhih...@gmail.com>;
>> *Cc:* "user@spark.apache.org"<user@spark.apache.org>;
>> *Subject: * Re: How to build Spark against the latest
>>
>>
>> What options should I use when running the make-distribution.sh script,
>>
>> I tried ./make-distribution.sh --hadoop.version 2.6.0 --with-yarn
>> -with-hive --with-tachyon --tgz
>> with nothing came out.
>>
>> Regards
>>
>> ------------------ Original ------------------
>> *From: * "guxiaobo1982";<guxiaobo1...@qq.com>;
>> *Send time:* Wednesday, Dec 24, 2014 6:52 PM
>> *To:* "Ted Yu"<yuzhih...@gmail.com>;
>> *Cc:* "user@spark.apache.org"<user@spark.apache.org>;
>> *Subject: * Re: How to build Spark against the latest
>>
>> Hi Ted,
>>      The reference command works, but where I can get the deployable
>> binaries?
>>
>> Xiaobo Gu
>>
>>
>>
>>
>> ------------------ Original ------------------
>> *From: * "Ted Yu";<yuzhih...@gmail.com>;
>> *Send time:* Wednesday, Dec 24, 2014 12:09 PM
>> *To:* ""<guxiaobo1...@qq.com>;
>> *Cc:* "user@spark.apache.org"<user@spark.apache.org>;
>> *Subject: * Re: How to build Spark against the latest
>>
>> See http://search-hadoop.com/m/JW1q5Cew0j
>>
>> On Tue, Dec 23, 2014 at 8:00 PM, guxiaobo1982 <guxiaobo1...@qq.com>
>> wrote:
>>
>>> Hi,
>>> The official pom.xml file only have profile for hadoop version 2.4 as
>>> the latest version, but I installed hadoop version 2.6.0 with ambari, how
>>> can I build spark against it, just using mvn -Dhadoop.version=2.6.0, or
>>> how to make a coresponding profile for it?
>>>
>>> Regards,
>>>
>>> Xiaobo
>>>
>>
>>
>

Reply via email to