Hi all,

I have a same error and it seems depending on Maven versions.

I tried building Spark using Maven with several versions on Jenkins.

+ Output of
"/Users/tomohiko/.jenkins/tools/hudson.tasks.Maven_MavenInstallation/mvn-3.3.3/bin/mvn
-version":

Apache Maven 3.3.3 (7994120775791599e205a5524ec3e0dfe41d4a06;
2015-04-22T20:57:37+09:00)
Maven home:
/Users/tomohiko/.jenkins/tools/hudson.tasks.Maven_MavenInstallation/mvn-3.3.3
Java version: 1.8.0, vendor: Oracle Corporation
Java home: /Library/Java/JavaVirtualMachines/jdk1.8.0.jdk/Contents/Home/jre
Default locale: en_US, platform encoding: UTF-8
OS name: "mac os x", version: "10.10.3", arch: "x86_64", family: "mac"

+ Jenkins Configuration:
Jenkins project type: Maven Project
Goals and options: -Phadoop-2.6 -DskipTests clean package

+ Maven versions and results:
3.3.3 -> infinite loop
3.3.1 -> infinite loop
3.2.5 -> SUCCESS


So do we prefer to build Spark with Maven 3.2.5?


On 4 July 2015 at 12:28, Andrew Or <and...@databricks.com> wrote:

> Thanks, I just tried it with 3.3.3 and I was able to reproduce it as well.
>
> 2015-07-03 18:51 GMT-07:00 Tarek Auel <tarek.a...@gmail.com>:
>
>> That's mine
>>
>> Apache Maven 3.3.3 (7994120775791599e205a5524ec3e0dfe41d4a06;
>> 2015-04-22T04:57:37-07:00)
>>
>> Maven home: /usr/local/Cellar/maven/3.3.3/libexec
>>
>> Java version: 1.8.0_45, vendor: Oracle Corporation
>>
>> Java home:
>> /Library/Java/JavaVirtualMachines/jdk1.8.0_45.jdk/Contents/Home/jre
>>
>> Default locale: en_US, platform encoding: UTF-8
>>
>> OS name: "mac os x", version: "10.10.3", arch: "x86_64", family: "mac"
>>
>> On Fri, Jul 3, 2015 at 6:32 PM Ted Yu <yuzhih...@gmail.com> wrote:
>>
>>> Here is mine:
>>>
>>> Apache Maven 3.3.1 (cab6659f9874fa96462afef40fcf6bc033d58c1c;
>>> 2015-03-13T13:10:27-07:00)
>>> Maven home: /home/hbase/apache-maven-3.3.1
>>> Java version: 1.8.0_45, vendor: Oracle Corporation
>>> Java home: /home/hbase/jdk1.8.0_45/jre
>>> Default locale: en_US, platform encoding: UTF-8
>>> OS name: "linux", version: "2.6.32-504.el6.x86_64", arch: "amd64",
>>> family: "unix"
>>>
>>> On Fri, Jul 3, 2015 at 6:05 PM, Andrew Or <and...@databricks.com> wrote:
>>>
>>>> @Tarek and Ted, what maven versions are you using?
>>>>
>>>> 2015-07-03 17:35 GMT-07:00 Krishna Sankar <ksanka...@gmail.com>:
>>>>
>>>>> Patrick,
>>>>>    I assume an RC3 will be out for folks like me to test the
>>>>> distribution. As usual, I will run the tests when you have a new
>>>>> distribution.
>>>>> Cheers
>>>>> <k/>
>>>>>
>>>>> On Fri, Jul 3, 2015 at 4:38 PM, Patrick Wendell <pwend...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Patch that added test-jar dependencies:
>>>>>> https://github.com/apache/spark/commit/bfe74b34
>>>>>>
>>>>>> Patch that originally disabled dependency reduced poms:
>>>>>>
>>>>>> https://github.com/apache/spark/commit/984ad60147c933f2d5a2040c87ae687c14eb1724
>>>>>>
>>>>>> Patch that reverted the disabling of dependency reduced poms:
>>>>>>
>>>>>> https://github.com/apache/spark/commit/bc51bcaea734fe64a90d007559e76f5ceebfea9e
>>>>>>
>>>>>> On Fri, Jul 3, 2015 at 4:36 PM, Patrick Wendell <pwend...@gmail.com>
>>>>>> wrote:
>>>>>> > Okay I did some forensics with Sean Owen. Some things about this
>>>>>> bug:
>>>>>> >
>>>>>> > 1. The underlying cause is that we added some code to make the tests
>>>>>> > of sub modules depend on the core tests. For unknown reasons this
>>>>>> > causes Spark to hit MSHADE-148 for *some* combinations of build
>>>>>> > profiles.
>>>>>> >
>>>>>> > 2. MSHADE-148 can be worked around by disabling building of
>>>>>> > "dependency reduced poms" because then the buggy code path is
>>>>>> > circumvented. Andrew Or did this in a patch on the 1.4 branch.
>>>>>> > However, that is not a tenable option for us because our *published*
>>>>>> > pom files require dependency reduction to substitute in the scala
>>>>>> > version correctly for the poms published to maven central.
>>>>>> >
>>>>>> > 3. As a result, Andrew Or reverted his patch recently, causing some
>>>>>> > package builds to start failing again (but publishing works now).
>>>>>> >
>>>>>> > 4. The reason this is not detected in our test harness or release
>>>>>> > build is that it is sensitive to the profiles enabled. The
>>>>>> combination
>>>>>> > of profiles we enable in the test harness and release builds do not
>>>>>> > trigger this bug.
>>>>>> >
>>>>>> > The best path I see forward right now is to do the following:
>>>>>> >
>>>>>> > 1. Disable creation of dependency reduced poms by default (this
>>>>>> > doesn't matter for people doing a package build) so typical users
>>>>>> > won't have this bug.
>>>>>> >
>>>>>> > 2. Add a profile that re-enables that setting.
>>>>>> >
>>>>>> > 3. Use the above profile when publishing release artifacts to maven
>>>>>> central.
>>>>>> >
>>>>>> > 4. Hope that we don't hit this bug for publishing.
>>>>>> >
>>>>>> > - Patrick
>>>>>> >
>>>>>> > On Fri, Jul 3, 2015 at 3:51 PM, Tarek Auel <tarek.a...@gmail.com>
>>>>>> wrote:
>>>>>> >> Doesn't change anything for me.
>>>>>> >>
>>>>>> >> On Fri, Jul 3, 2015 at 3:45 PM Patrick Wendell <pwend...@gmail.com>
>>>>>> wrote:
>>>>>> >>>
>>>>>> >>> Can you try using the built in maven "build/mvn..."? All of our
>>>>>> builds
>>>>>> >>> are passing on Jenkins so I wonder if it's a maven version issue:
>>>>>> >>>
>>>>>> >>> https://amplab.cs.berkeley.edu/jenkins/view/Spark-QA-Compile/
>>>>>> >>>
>>>>>> >>> - Patrick
>>>>>> >>>
>>>>>> >>> On Fri, Jul 3, 2015 at 3:14 PM, Ted Yu <yuzhih...@gmail.com>
>>>>>> wrote:
>>>>>> >>> > Please take a look at SPARK-8781
>>>>>> >>> > (https://github.com/apache/spark/pull/7193)
>>>>>> >>> >
>>>>>> >>> > Cheers
>>>>>> >>> >
>>>>>> >>> > On Fri, Jul 3, 2015 at 3:05 PM, Tarek Auel <
>>>>>> tarek.a...@gmail.com> wrote:
>>>>>> >>> >>
>>>>>> >>> >> I found a solution, there might be a better one.
>>>>>> >>> >>
>>>>>> >>> >> https://github.com/apache/spark/pull/7217
>>>>>> >>> >>
>>>>>> >>> >> On Fri, Jul 3, 2015 at 2:28 PM Robin East <
>>>>>> robin.e...@xense.co.uk>
>>>>>> >>> >> wrote:
>>>>>> >>> >>>
>>>>>> >>> >>> Yes me too
>>>>>> >>> >>>
>>>>>> >>> >>> On 3 Jul 2015, at 22:21, Ted Yu <yuzhih...@gmail.com> wrote:
>>>>>> >>> >>>
>>>>>> >>> >>> This is what I got (the last line was repeated non-stop):
>>>>>> >>> >>>
>>>>>> >>> >>> [INFO] Replacing original artifact with shaded artifact.
>>>>>> >>> >>> [INFO] Replacing
>>>>>> >>> >>>
>>>>>> /home/hbase/spark/bagel/target/spark-bagel_2.10-1.5.0-SNAPSHOT.jar
>>>>>> >>> >>> with
>>>>>> >>> >>>
>>>>>> >>> >>>
>>>>>> /home/hbase/spark/bagel/target/spark-bagel_2.10-1.5.0-SNAPSHOT-shaded.jar
>>>>>> >>> >>> [INFO] Dependency-reduced POM written at:
>>>>>> >>> >>> /home/hbase/spark/bagel/dependency-reduced-pom.xml
>>>>>> >>> >>> [INFO] Dependency-reduced POM written at:
>>>>>> >>> >>> /home/hbase/spark/bagel/dependency-reduced-pom.xml
>>>>>> >>> >>>
>>>>>> >>> >>> On Fri, Jul 3, 2015 at 1:13 PM, Tarek Auel <
>>>>>> tarek.a...@gmail.com>
>>>>>> >>> >>> wrote:
>>>>>> >>> >>>>
>>>>>> >>> >>>> Hi all,
>>>>>> >>> >>>>
>>>>>> >>> >>>> I am trying to build the master, but it stucks and prints
>>>>>> >>> >>>>
>>>>>> >>> >>>> [INFO] Dependency-reduced POM written at:
>>>>>> >>> >>>> /Users/tarek/test/spark/bagel/dependency-reduced-pom.xml
>>>>>> >>> >>>>
>>>>>> >>> >>>> build command:  mvn -DskipTests clean package
>>>>>> >>> >>>>
>>>>>> >>> >>>> Do others have the same issue?
>>>>>> >>> >>>>
>>>>>> >>> >>>> Regards,
>>>>>> >>> >>>> Tarek
>>>>>> >>> >>>
>>>>>> >>> >>>
>>>>>> >>> >>>
>>>>>> >>> >
>>>>>>
>>>>>> ---------------------------------------------------------------------
>>>>>> To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
>>>>>> For additional commands, e-mail: dev-h...@spark.apache.org
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>


-- 
class Cocoatomo:
    name = 'cocoatomo'
    email_address = 'cocoatom...@gmail.com'
    twitter_id = '@cocoatomo'

Reply via email to