Seems the build and directory structure in dist is similar to the .gz file
downloaded from the
downloads page. Can the dist directory be used as is?

On Tue, Oct 27, 2015 at 4:03 PM, Kayode Odeyemi <drey...@gmail.com> wrote:

> Ted, I switched to this:
>
> ./make-distribution.sh --name spark-latest --tgz -Dhadoop.version=2.6.0
> -Phadoop-2.6 -Phive -Phive-thriftserver -Pyarn -DskipTests clean package -U
>
> Same error. No .gz file. Here's the bottom output log:
>
> + rm -rf /home/emperor/javaprojects/spark/dist
> + mkdir -p /home/emperor/javaprojects/spark/dist/lib
> + echo 'Spark [WARNING] See
> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin (git revision
> 3689beb) built for Hadoop [WARNING] See
> http://docs.codehaus.org/display/MAVENUSER/Shade+Pl
> + echo 'Build flags: -Dhadoop.version=2.6.0' -Phadoop-2.6 -Phive
> -Phive-thriftserver -Pyarn -DskipTests clean package -U
> + cp
> /home/emperor/javaprojects/spark/assembly/target/scala-2.10/spark-assembly-1.6.0-SNAPSHOT-hadoop2.6.0.jar
> /home/emperor/javaprojects/spark/dist/lib/
> + cp
> /home/emperor/javaprojects/spark/examples/target/scala-2.10/spark-examples-1.6.0-SNAPSHOT-hadoop2.6.0.jar
> /home/emperor/javaprojects/spark/dist/lib/
> + cp
> /home/emperor/javaprojects/spark/network/yarn/target/scala-2.10/spark-1.6.0-SNAPSHOT-yarn-shuffle.jar
> /home/emperor/javaprojects/spark/dist/lib/
> + mkdir -p /home/emperor/javaprojects/spark/dist/examples/src/main
> + cp -r /home/emperor/javaprojects/spark/examples/src/main
> /home/emperor/javaprojects/spark/dist/examples/src/
> + '[' 1 == 1 ']'
> + cp
> /home/emperor/javaprojects/spark/lib_managed/jars/datanucleus-api-jdo-3.2.6.jar
> /home/emperor/javaprojects/spark/lib_managed/jars/datanucleus-core-3.2.10.jar
> /home/emperor/javaprojects
> ed/jars/datanucleus-rdbms-3.2.9.jar
> /home/emperor/javaprojects/spark/dist/lib/
> + cp /home/emperor/javaprojects/spark/LICENSE
> /home/emperor/javaprojects/spark/dist
> + cp -r /home/emperor/javaprojects/spark/licenses
> /home/emperor/javaprojects/spark/dist
> + cp /home/emperor/javaprojects/spark/NOTICE
> /home/emperor/javaprojects/spark/dist
> + '[' -e /home/emperor/javaprojects/spark/CHANGES.txt ']'
> + cp -r /home/emperor/javaprojects/spark/data
> /home/emperor/javaprojects/spark/dist
> + mkdir /home/emperor/javaprojects/spark/dist/conf
> + cp /home/emperor/javaprojects/spark/conf/docker.properties.template
> /home/emperor/javaprojects/spark/conf/fairscheduler.xml.template
> /home/emperor/javaprojects/spark/conf/log4j.properties
> emperor/javaprojects/spark/conf/metrics.properties.template
> /home/emperor/javaprojects/spark/conf/slaves.template
> /home/emperor/javaprojects/spark/conf/spark-defaults.conf.template /home/em
> ts/spark/conf/spark-env.sh.template
> /home/emperor/javaprojects/spark/dist/conf
> + cp /home/emperor/javaprojects/spark/README.md
> /home/emperor/javaprojects/spark/dist
> + cp -r /home/emperor/javaprojects/spark/bin
> /home/emperor/javaprojects/spark/dist
> + cp -r /home/emperor/javaprojects/spark/python
> /home/emperor/javaprojects/spark/dist
> + cp -r /home/emperor/javaprojects/spark/sbin
> /home/emperor/javaprojects/spark/dist
> + cp -r /home/emperor/javaprojects/spark/ec2
> /home/emperor/javaprojects/spark/dist
> + '[' -d /home/emperor/javaprojects/spark/R/lib/SparkR ']'
> + '[' false == true ']'
> + '[' true == true ']'
> + TARDIR_NAME='spark-[WARNING] See
> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest'
> + TARDIR='/home/emperor/javaprojects/spark/spark-[WARNING] See
> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest'
> + rm -rf '/home/emperor/javaprojects/spark/spark-[WARNING] See
> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest'
> + cp -r /home/emperor/javaprojects/spark/dist
> '/home/emperor/javaprojects/spark/spark-[WARNING] See
> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest'
> cp: cannot create directory
> `/home/emperor/javaprojects/spark/spark-[WARNING] See
> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest':
> No such file or directory
>
>
> On Tue, Oct 27, 2015 at 2:14 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>
>> Can you try the same command shown in the pull request ?
>>
>> Thanks
>>
>> On Oct 27, 2015, at 12:40 AM, Kayode Odeyemi <drey...@gmail.com> wrote:
>>
>> Thank you.
>>
>> But I'm getting same warnings and it's still preventing the archive from
>> being generated.
>>
>> I've ran this on both OSX Lion and Ubuntu 12. Same error. No .gz file
>>
>> On Mon, Oct 26, 2015 at 9:10 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>
>>> Looks like '-Pyarn' was missing in your command.
>>>
>>> On Mon, Oct 26, 2015 at 12:06 PM, Kayode Odeyemi <drey...@gmail.com>
>>> wrote:
>>>
>>>> I used this command which is synonymous to what you have:
>>>>
>>>> ./make-distribution.sh --name spark-latest --tgz --mvn mvn
>>>> -Dhadoop.version=2.6.0 -Phadoop-2.6 -Phive -Phive-thriftserver -DskipTests
>>>> clean package -U
>>>>
>>>> But I still see WARNINGS like this in the output and no .gz file
>>>> created:
>>>>
>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=26/.part-r-00005.gz.parquet.crc:
>>>> No such file or directory
>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=26/part-r-00005.gz.parquet:
>>>> No such file or directory
>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9:
>>>> No such file or directory
>>>> cp:
>>>> /usr/local/spark-latest/dist/python/test_support/sql/parquet_partitioned/year=2015/month=9:
>>>> unable to copy extended attributes to
>>>> /usr/local/spark-latest/spark-[WARNING] See
>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9:
>>>> No such file or directory
>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1:
>>>> No such file or directory
>>>> cp:
>>>> /usr/local/spark-latest/dist/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1:
>>>> unable to copy extended attributes to
>>>> /usr/local/spark-latest/spark-[WARNING] See
>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1:
>>>> No such file or directory
>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1/.part-r-00007.gz.parquet.crc:
>>>> No such file or directory
>>>>
>>>> On Mon, Oct 26, 2015 at 8:58 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>>>>
>>>>> If you use the command shown in:
>>>>> https://github.com/apache/spark/pull/9281
>>>>>
>>>>> You should have got the following:
>>>>>
>>>>>
>>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2014/month=9/day=1/part-r-00008.gz.parquet
>>>>>
>>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=9/day=1/part-r-00007.gz.parquet
>>>>>
>>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=25/part-r-00004.gz.parquet
>>>>>
>>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=25/part-r-00002.gz.parquet
>>>>>
>>>>> ./dist/python/test_support/sql/parquet_partitioned/year=2015/month=10/day=26/part-r-00005.gz.parquet
>>>>>
>>>>> On Mon, Oct 26, 2015 at 11:47 AM, Kayode Odeyemi <drey...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> I see a lot of stuffs like this after the a successful maven build:
>>>>>>
>>>>>> cp: /usr/local/spark-latest/spark-[WARNING] See
>>>>>> http://docs.codehaus.org/display/MAVENUSER/Shade+Plugin-bin-spark-latest/python/test_support/sql/parquet_partitioned/year=2014/month=9/day=1/
>>>>>> part-r-00008.gz.parquet: No such file or directory
>>>>>>
>>>>>> Seems it fails when it tries to package the build as an archive.
>>>>>>
>>>>>> I'm using the latest code on github master.
>>>>>>
>>>>>> Any ideas please?
>>>>>>
>>>>>> On Mon, Oct 26, 2015 at 6:20 PM, Yana Kadiyska <
>>>>>> yana.kadiy...@gmail.com> wrote:
>>>>>>
>>>>>>> In 1.4 ./make_distribution produces a .tgz file in the root
>>>>>>> directory (same directory that make_distribution is in)
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Oct 26, 2015 at 8:46 AM, Kayode Odeyemi <drey...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> The ./make_distribution task completed. However, I can't seem to
>>>>>>>> locate the
>>>>>>>> .tar.gz file.
>>>>>>>>
>>>>>>>> Where does Spark save this? or should I just work with the dist
>>>>>>>> directory?
>>>>>>>>
>>>>>>>> On Fri, Oct 23, 2015 at 4:23 PM, Kayode Odeyemi <drey...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> I saw this when I tested manually (without ./make-distribution)
>>>>>>>>>
>>>>>>>>> Detected Maven Version: 3.2.2 is not in the allowed range 3.3.3.
>>>>>>>>>
>>>>>>>>> So I simply upgraded maven to 3.3.3.
>>>>>>>>>
>>>>>>>>> Resolved. Thanks
>>>>>>>>>
>>>>>>>>> On Fri, Oct 23, 2015 at 3:17 PM, Sean Owen <so...@cloudera.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> This doesn't show the actual error output from Maven. I have a
>>>>>>>>>> strong
>>>>>>>>>> guess that you haven't set MAVEN_OPTS to increase the memory
>>>>>>>>>> Maven can
>>>>>>>>>> use.
>>>>>>>>>>
>>>>>>>>>> On Fri, Oct 23, 2015 at 6:14 AM, Kayode Odeyemi <
>>>>>>>>>> drey...@gmail.com> wrote:
>>>>>>>>>> > Hi,
>>>>>>>>>> >
>>>>>>>>>> > I can't seem to get a successful maven build. Please see
>>>>>>>>>> command output
>>>>>>>>>> > below:
>>>>>>>>>> >
>>>>>>>>>> > bash-3.2$ ./make-distribution.sh --name spark-latest --tgz
>>>>>>>>>> --mvn mvn
>>>>>>>>>> > -Dhadoop.version=2.7.0 -Phadoop-2.7 -Phive -Phive-thriftserver
>>>>>>>>>> -DskipTests
>>>>>>>>>> > clean package
>>>>>>>>>> > +++ dirname ./make-distribution.sh
>>>>>>>>>> > ++ cd .
>>>>>>>>>> > ++ pwd
>>>>>>>>>> > + SPARK_HOME=/usr/local/spark-latest
>>>>>>>>>> > + DISTDIR=/usr/local/spark-latest/dist
>>>>>>>>>> > + SPARK_TACHYON=false
>>>>>>>>>> > + TACHYON_VERSION=0.7.1
>>>>>>>>>> > + TACHYON_TGZ=tachyon-0.7.1-bin.tar.gz
>>>>>>>>>> > +
>>>>>>>>>> > TACHYON_URL=
>>>>>>>>>> https://github.com/amplab/tachyon/releases/download/v0.7.1/tachyon-0.7.1-bin.tar.gz
>>>>>>>>>> > + MAKE_TGZ=false
>>>>>>>>>> > + NAME=none
>>>>>>>>>> > + MVN=/usr/local/spark-latest/build/mvn
>>>>>>>>>> > + ((  12  ))
>>>>>>>>>> > + case $1 in
>>>>>>>>>> > + NAME=spark-latest
>>>>>>>>>> > + shift
>>>>>>>>>> > + shift
>>>>>>>>>> > + ((  10  ))
>>>>>>>>>> > + case $1 in
>>>>>>>>>> > + MAKE_TGZ=true
>>>>>>>>>> > + shift
>>>>>>>>>> > + ((  9  ))
>>>>>>>>>> > + case $1 in
>>>>>>>>>> > + MVN=mvn
>>>>>>>>>> > + shift
>>>>>>>>>> > + shift
>>>>>>>>>> > + ((  7  ))
>>>>>>>>>> > + case $1 in
>>>>>>>>>> > + break
>>>>>>>>>> > + '[' -z
>>>>>>>>>> /Library/Java/JavaVirtualMachines/jdk1.8.0_20.jdk/Contents/Home ']'
>>>>>>>>>> > + '[' -z
>>>>>>>>>> /Library/Java/JavaVirtualMachines/jdk1.8.0_20.jdk/Contents/Home ']'
>>>>>>>>>> > ++ command -v git
>>>>>>>>>> > + '[' /usr/bin/git ']'
>>>>>>>>>> > ++ git rev-parse --short HEAD
>>>>>>>>>> > + GITREV=487d409
>>>>>>>>>> > + '[' '!' -z 487d409 ']'
>>>>>>>>>> > + GITREVSTRING=' (git revision 487d409)'
>>>>>>>>>> > + unset GITREV
>>>>>>>>>> > ++ command -v mvn
>>>>>>>>>> > + '[' '!' /usr/bin/mvn ']'
>>>>>>>>>> > ++ mvn help:evaluate -Dexpression=project.version
>>>>>>>>>> -Dhadoop.version=2.7.0
>>>>>>>>>> > -Phadoop-2.7 -Phive -Phive-thriftserver -DskipTests clean
>>>>>>>>>> package
>>>>>>>>>> > ++ grep -v INFO
>>>>>>>>>> > ++ tail -n 1
>>>>>>>>>> > + VERSION='[ERROR] [Help 1]
>>>>>>>>>> >
>>>>>>>>>> http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
>>>>>>>>>> '
>>>>>>>>>> >
>>>>>>>>>> > Same output error with JDK 7
>>>>>>>>>> >
>>>>>>>>>> > Appreciate your help.
>>>>>>>>>> >
>>>>>>>>>> >
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to