Ted, I switched to this:

./make-distribution.sh --name spark-latest --tgz -Dhadoop.version=2.6.0
-Phadoop-2.6 -Phive -Phive-thriftserver -Pyarn -DskipTests clean package -U

Same error. No .gz file. Here's the bottom output log:

+ rm -rf /home/emperor/javaprojects/spark/dist
+ mkdir -p /home/emperor/javaprojects/spark/dist/lib
+ echo 'Spark [WARNING] See
http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin (git revision
3689beb) built for Hadoop [WARNING] See
http://docs.codehaus.org/display/MAVENUSER/Shade+Pl
+ echo 'Build flags: -Dhadoop.version=2.6.0' -Phadoop-2.6 -Phive
-Phive-thriftserver -Pyarn -DskipTests clean package -U
+ cp
/home/emperor/javaprojects/spark/assembly/target/scala-2.10/spark-assembly-1.6.0-SNAPSHOT-hadoop2.6.0.jar
/home/emperor/javaprojects/spark/dist/lib/
+ cp
/home/emperor/javaprojects/spark/examples/target/scala-2.10/spark-examples-1.6.0-SNAPSHOT-hadoop2.6.0.jar
/home/emperor/javaprojects/spark/dist/lib/
+ cp
/home/emperor/javaprojects/spark/network/yarn/target/scala-2.10/spark-1.6.0-SNAPSHOT-yarn-shuffle.jar
/home/emperor/javaprojects/spark/dist/lib/
+ mkdir -p /home/emperor/javaprojects/spark/dist/examples/src/main
+ cp -r /home/emperor/javaprojects/spark/examples/src/main
/home/emperor/javaprojects/spark/dist/examples/src/
+ '[' 1 == 1 ']'
+ cp
/home/emperor/javaprojects/spark/lib_managed/jars/datanucleus-api-jdo-3.2.6.jar
/home/emperor/javaprojects/spark/lib_managed/jars/datanucleus-core-3.2.10.jar
/home/emperor/javaprojects
ed/jars/datanucleus-rdbms-3.2.9.jar
/home/emperor/javaprojects/spark/dist/lib/
+ cp /home/emperor/javaprojects/spark/LICENSE
/home/emperor/javaprojects/spark/dist
+ cp -r /home/emperor/javaprojects/spark/licenses
/home/emperor/javaprojects/spark/dist
+ cp /home/emperor/javaprojects/spark/NOTICE
/home/emperor/javaprojects/spark/dist
+ '[' -e /home/emperor/javaprojects/spark/CHANGES.txt ']'
+ cp -r /home/emperor/javaprojects/spark/data
/home/emperor/javaprojects/spark/dist
+ mkdir /home/emperor/javaprojects/spark/dist/conf
+ cp /home/emperor/javaprojects/spark/conf/docker.properties.template
/home/emperor/javaprojects/spark/conf/fairscheduler.xml.template
/home/emperor/javaprojects/spark/conf/log4j.properties
emperor/javaprojects/spark/conf/metrics.properties.template
/home/emperor/javaprojects/spark/conf/slaves.template
/home/emperor/javaprojects/spark/conf/spark-defaults.conf.template /home/em
ts/spark/conf/spark-env.sh.template
/home/emperor/javaprojects/spark/dist/conf
+ cp /home/emperor/javaprojects/spark/README.md
/home/emperor/javaprojects/spark/dist
+ cp -r /home/emperor/javaprojects/spark/bin
/home/emperor/javaprojects/spark/dist
+ cp -r /home/emperor/javaprojects/spark/python
/home/emperor/javaprojects/spark/dist
+ cp -r /home/emperor/javaprojects/spark/sbin
/home/emperor/javaprojects/spark/dist
+ cp -r /home/emperor/javaprojects/spark/ec2
/home/emperor/javaprojects/spark/dist
+ '[' -d /home/emperor/javaprojects/spark/R/lib/SparkR ']'
+ '[' false == true ']'
+ '[' true == true ']'
+ TARDIR_NAME='spark-[WARNING] See
http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest'
+ TARDIR='/home/emperor/javaprojects/spark/spark-[WARNING] See
http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest'
+ rm -rf '/home/emperor/javaprojects/spark/spark-[WARNING] See
http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest'
+ cp -r /home/emperor/javaprojects/spark/dist
'/home/emperor/javaprojects/spark/spark-[WARNING] See
http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest'
cp: cannot create directory
`/home/emperor/javaprojects/spark/spark-[WARNING] See
http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest':
No such file or directory


On Tue, Oct 27, 2015 at 2:14 PM, Ted Yu <yuzhih...@gmail.com> wrote:

> Can you try the same command shown in the pull request ?
>
> Thanks
>
> On Oct 27, 2015, at 12:40 AM, Kayode Odeyemi <drey...@gmail.com> wrote:
>
> Thank you.
>
> But I'm getting same warnings and it's still preventing the archive from
> being generated.
>
> I've ran this on both OSX Lion and Ubuntu 12. Same error. No .gz file
>
> On Mon, Oct 26, 2015 at 9:10 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>
>> Looks like '-Pyarn' was missing in your command.
>>
>> On Mon, Oct 26, 2015 at 12:06 PM, Kayode Odeyemi <drey...@gmail.com>
>> wrote:
>>
>>> I used this command which is synonymous to what you have:
>>>
>>> ./make-distribution.sh --name spark-latest --tgz --mvn mvn
>>> -Dhadoop.version=2.6.0 -Phadoop-2.6 -Phive -Phive-thriftserver -DskipTests
>>> clean package -U
>>>
>>> But I still see WARNINGS like this in the output and no .gz file created:
>>>
>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=26/.part-r-00005.gz.parquet.crc:
>>> No such file or directory
>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=26/part-r-00005.gz.parquet:
>>> No such file or directory
>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9:
>>> No such file or directory
>>> cp:
>>> /usr/local/spark-latest/dist/python/test_support/sql/parquet_partitioned/year=2015/month=9:
>>> unable to copy extended attributes to
>>> /usr/local/spark-latest/spark-[WARNING] See
>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9:
>>> No such file or directory
>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1:
>>> No such file or directory
>>> cp:
>>> /usr/local/spark-latest/dist/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1:
>>> unable to copy extended attributes to
>>> /usr/local/spark-latest/spark-[WARNING] See
>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1:
>>> No such file or directory
>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1/.part-r-00007.gz.parquet.crc:
>>> No such file or directory
>>>
>>> On Mon, Oct 26, 2015 at 8:58 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>>
>>>> If you use the command shown in:
>>>> https://github.com/apache/spark/pull/9281
>>>>
>>>> You should have got the following:
>>>>
>>>>
>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2014/month=9/day=1/part-r-00008.gz.parquet
>>>>
>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1/part-r-00007.gz.parquet
>>>>
>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=25/part-r-00004.gz.parquet
>>>>
>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=25/part-r-00002.gz.parquet
>>>>
>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=26/part-r-00005.gz.parquet
>>>>
>>>> On Mon, Oct 26, 2015 at 11:47 AM, Kayode Odeyemi <drey...@gmail.com>
>>>> wrote:
>>>>
>>>>> I see a lot of stuffs like this after the a successful maven build:
>>>>>
>>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2014/month=9/day=1/
>>>>> part-r-00008.gz.parquet: No such file or directory
>>>>>
>>>>> Seems it fails when it tries to package the build as an archive.
>>>>>
>>>>> I'm using the latest code on github master.
>>>>>
>>>>> Any ideas please?
>>>>>
>>>>> On Mon, Oct 26, 2015 at 6:20 PM, Yana Kadiyska <
>>>>> yana.kadiy...@gmail.com> wrote:
>>>>>
>>>>>> In 1.4 ./make_distribution produces a .tgz file in the root
>>>>>> directory (same directory that make_distribution is in)
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Mon, Oct 26, 2015 at 8:46 AM, Kayode Odeyemi <drey...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> The ./make_distribution task completed. However, I can't seem to
>>>>>>> locate the
>>>>>>> .tar.gz file.
>>>>>>>
>>>>>>> Where does Spark save this? or should I just work with the dist
>>>>>>> directory?
>>>>>>>
>>>>>>> On Fri, Oct 23, 2015 at 4:23 PM, Kayode Odeyemi <drey...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> I saw this when I tested manually (without ./make-distribution)
>>>>>>>>
>>>>>>>> Detected Maven Version: 3.2.2 is not in the allowed range 3.3.3.
>>>>>>>>
>>>>>>>> So I simply upgraded maven to 3.3.3.
>>>>>>>>
>>>>>>>> Resolved. Thanks
>>>>>>>>
>>>>>>>> On Fri, Oct 23, 2015 at 3:17 PM, Sean Owen <so...@cloudera.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> This doesn't show the actual error output from Maven. I have a
>>>>>>>>> strong
>>>>>>>>> guess that you haven't set MAVEN_OPTS to increase the memory Maven
>>>>>>>>> can
>>>>>>>>> use.
>>>>>>>>>
>>>>>>>>> On Fri, Oct 23, 2015 at 6:14 AM, Kayode Odeyemi <drey...@gmail.com>
>>>>>>>>> wrote:
>>>>>>>>> > Hi,
>>>>>>>>> >
>>>>>>>>> > I can't seem to get a successful maven build. Please see command
>>>>>>>>> output
>>>>>>>>> > below:
>>>>>>>>> >
>>>>>>>>> > bash-3.2$ ./make-distribution.sh --name spark-latest --tgz --mvn
>>>>>>>>> mvn
>>>>>>>>> > -Dhadoop.version=2.7.0 -Phadoop-2.7 -Phive -Phive-thriftserver
>>>>>>>>> -DskipTests
>>>>>>>>> > clean package
>>>>>>>>> > +++ dirname ./make-distribution.sh
>>>>>>>>> > ++ cd .
>>>>>>>>> > ++ pwd
>>>>>>>>> > + SPARK_HOME=/usr/local/spark-latest
>>>>>>>>> > + DISTDIR=/usr/local/spark-latest/dist
>>>>>>>>> > + SPARK_TACHYON=false
>>>>>>>>> > + TACHYON_VERSION=0.7.1
>>>>>>>>> > + TACHYON_TGZ=tachyon-0.7.1-bin.tar.gz
>>>>>>>>> > +
>>>>>>>>> > TACHYON_URL=
>>>>>>>>> https://github.com/amplab/tachyon/releases/download/v0.7.1/tachyon-0.7.1-bin.tar.gz
>>>>>>>>> > + MAKE_TGZ=false
>>>>>>>>> > + NAME=none
>>>>>>>>> > + MVN=/usr/local/spark-latest/build/mvn
>>>>>>>>> > + ((  12  ))
>>>>>>>>> > + case $1 in
>>>>>>>>> > + NAME=spark-latest
>>>>>>>>> > + shift
>>>>>>>>> > + shift
>>>>>>>>> > + ((  10  ))
>>>>>>>>> > + case $1 in
>>>>>>>>> > + MAKE_TGZ=true
>>>>>>>>> > + shift
>>>>>>>>> > + ((  9  ))
>>>>>>>>> > + case $1 in
>>>>>>>>> > + MVN=mvn
>>>>>>>>> > + shift
>>>>>>>>> > + shift
>>>>>>>>> > + ((  7  ))
>>>>>>>>> > + case $1 in
>>>>>>>>> > + break
>>>>>>>>> > + '[' -z
>>>>>>>>> /Library/Java/JavaVirtualMachines/jdk1.8.0_20.jdk/Contents/Home ']'
>>>>>>>>> > + '[' -z
>>>>>>>>> /Library/Java/JavaVirtualMachines/jdk1.8.0_20.jdk/Contents/Home ']'
>>>>>>>>> > ++ command -v git
>>>>>>>>> > + '[' /usr/bin/git ']'
>>>>>>>>> > ++ git rev-parse --short HEAD
>>>>>>>>> > + GITREV=487d409
>>>>>>>>> > + '[' '!' -z 487d409 ']'
>>>>>>>>> > + GITREVSTRING=' (git revision 487d409)'
>>>>>>>>> > + unset GITREV
>>>>>>>>> > ++ command -v mvn
>>>>>>>>> > + '[' '!' /usr/bin/mvn ']'
>>>>>>>>> > ++ mvn help:evaluate -Dexpression=project.version
>>>>>>>>> -Dhadoop.version=2.7.0
>>>>>>>>> > -Phadoop-2.7 -Phive -Phive-thriftserver -DskipTests clean package
>>>>>>>>> > ++ grep -v INFO
>>>>>>>>> > ++ tail -n 1
>>>>>>>>> > + VERSION='[ERROR] [Help 1]
>>>>>>>>> >
>>>>>>>>> http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
>>>>>>>>> '
>>>>>>>>> >
>>>>>>>>> > Same output error with JDK 7
>>>>>>>>> >
>>>>>>>>> > Appreciate your help.
>>>>>>>>> >
>>>>>>>>> >
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to