I just tried the hdfs build on Apache Jenkins. Its still failing:

https://builds.apache.org/job/Hadoop-Hdfs-trunk/752/console

Looks like something you are noticing Todd.

Also, is https://issues.apache.org/jira/browse/HDFS-2261 also an issue for the 
builds? 

thanks
mahadev
On Aug 15, 2011, at 4:58 PM, Todd Lipcon wrote:

> I'm having some related issues locally... seems every time the Hudson
> publishes a new build, my local build breaks with something like this:
> 
> 
> todd@todd-w510:~/git/hadoop-common/hdfs$ ant clean test
> Buildfile: /home/todd/git/hadoop-common/hdfs/build.xml
> 
> clean-contrib:
> 
> clean:
> 
> clean:
>     [echo] contrib: fuse-dfs
> 
> clean-fi:
> 
> clean-sign:
> 
> clean:
>   [delete] Deleting directory /home/todd/git/hadoop-common/hdfs/build
> 
> ivy-download:
>      [get] Getting:
> http://repo2.maven.org/maven2/org/apache/ivy/ivy/2.2.0-rc1/ivy-2.2.0-rc1.jar
>      [get] To: /home/todd/git/hadoop-common/hdfs/ivy/ivy-2.2.0-rc1.jar
>      [get] Not modified - so not downloaded
> 
> ivy-init-dirs:
>    [mkdir] Created dir: /home/todd/git/hadoop-common/hdfs/build/ivy
>    [mkdir] Created dir: /home/todd/git/hadoop-common/hdfs/build/ivy/lib
>    [mkdir] Created dir: /home/todd/git/hadoop-common/hdfs/build/ivy/report
>    [mkdir] Created dir: /home/todd/git/hadoop-common/hdfs/build/ivy/maven
> 
> ivy-probe-antlib:
> 
> ivy-init-antlib:
> 
> ivy-init:
> [ivy:configure] :: Ivy 2.2.0-rc1 - 20100629224905 ::
> http://ant.apache.org/ivy/ ::
> [ivy:configure] :: loading settings :: file =
> /home/todd/git/hadoop-common/hdfs/ivy/ivysettings.xml
> 
> ivy-resolve-common:
> [ivy:resolve] downloading
> https://repository.apache.org/content/repositories/snapshots/org/apache/hadoop/hadoop-annotations/0.23.0-SNAPSHOT/hadoop-annotations-0.23.0-20110808.090045-15.jar
> ...
> [ivy:resolve] ... (14kB)
> [ivy:resolve] .. (0kB)
> [ivy:resolve]   [SUCCESSFUL ]
> org.apache.hadoop#hadoop-annotations;0.23.0-SNAPSHOT!hadoop-annotations.jar
> (458ms)
> [ivy:resolve] downloading
> https://repository.apache.org/content/repositories/snapshots/org/apache/hadoop/hadoop-common/0.23.0-SNAPSHOT/hadoop-common-0.23.0-20110815.225725-267.jar
> ...
> [ivy:resolve] 
> .....................................................................................................................
> [ivy:resolve] 
> ....................................................................................................
> (1667kB)
> [ivy:resolve] .. (0kB)
> [ivy:resolve]   [SUCCESSFUL ]
> org.apache.hadoop#hadoop-common;0.23.0-SNAPSHOT!hadoop-common.jar
> (2796ms)
> 
> ivy-retrieve-common:
> [ivy:cachepath] DEPRECATED: 'ivy.conf.file' is deprecated, use
> 'ivy.settings.file' instead
> [ivy:cachepath] :: loading settings :: file =
> /home/todd/git/hadoop-common/hdfs/ivy/ivysettings.xml
> 
> ivy-resolve-hdfs:
> [ivy:resolve] downloading
> https://repository.apache.org/content/repositories/snapshots/org/apache/hadoop/hadoop-annotations/0.23.0-SNAPSHOT/hadoop-annotations-0.23.0-20110808.090045-15.jar
> ...
> [ivy:resolve] ... (14kB)
> [ivy:resolve] .. (0kB)
> [ivy:resolve]   [SUCCESSFUL ]
> org.apache.hadoop#hadoop-annotations;0.23.0-SNAPSHOT!hadoop-annotations.jar
> (193ms)
> 
> ivy-retrieve-hdfs:
> 
> ivy-resolve-test:
> 
> ivy-retrieve-test:
> 
> init:
>    [mkdir] Created dir: /home/todd/git/hadoop-common/hdfs/build/classes
>    [mkdir] Created dir: /home/todd/git/hadoop-common/hdfs/build/src
>    [mkdir] Created dir:
> /home/todd/git/hadoop-common/hdfs/build/web/webapps/hdfs/WEB-INF
>    [mkdir] Created dir:
> /home/todd/git/hadoop-common/hdfs/build/web/webapps/datanode/WEB-INF
>    [mkdir] Created dir:
> /home/todd/git/hadoop-common/hdfs/build/web/webapps/secondary/WEB-INF
>    [mkdir] Created dir: /home/todd/git/hadoop-common/hdfs/build/ant
>    [mkdir] Created dir: /home/todd/git/hadoop-common/hdfs/build/c++
>    [mkdir] Created dir: /home/todd/git/hadoop-common/hdfs/build/test
>    [mkdir] Created dir:
> /home/todd/git/hadoop-common/hdfs/build/test/hdfs/classes
>    [mkdir] Created dir: /home/todd/git/hadoop-common/hdfs/build/test/extraconf
>    [touch] Creating /tmp/null1539574669
>   [delete] Deleting: /tmp/null1539574669
>     [copy] Copying 6 files to
> /home/todd/git/hadoop-common/hdfs/build/web/webapps
>     [copy] Copying 1 file to /home/todd/git/hadoop-common/hdfs/conf
>     [copy] Copying
> /home/todd/git/hadoop-common/hdfs/conf/hdfs-site.xml.template to
> /home/todd/git/hadoop-common/hdfs/conf/hdfs-site.xml
>    [mkdir] Created dir: /home/todd/git/hadoop-common/hdfs/build/test/conf
>     [copy] Copying 1 file to /home/todd/git/hadoop-common/hdfs/build/test/conf
>     [copy] Copying
> /home/todd/git/hadoop-common/hdfs/conf/hdfs-site.xml.template to
> /home/todd/git/hadoop-common/hdfs/build/test/conf/hdfs-site.xml
>     [copy] Copying 2 files to 
> /home/todd/git/hadoop-common/hdfs/build/test/conf
>     [copy] Copying
> /home/todd/git/hadoop-common/hdfs/conf/hadoop-metrics2.properties to
> /home/todd/git/hadoop-common/hdfs/build/test/conf/hadoop-metrics2.properties
>     [copy] Copying
> /home/todd/git/hadoop-common/hdfs/conf/log4j.properties to
> /home/todd/git/hadoop-common/hdfs/build/test/conf/log4j.properties
> 
> check-libhdfs-makefile:
> 
> create-libhdfs-makefile:
> 
> compile-c++-libhdfs:
> 
> clover.setup:
> 
> clover.info:
>     [echo]
>     [echo]      Clover not found. Code coverage reports disabled.
>     [echo]
> 
> clover:
> 
> compile-hdfs-classes:
>    [javac] /home/todd/git/hadoop-common/hdfs/build.xml:370: warning:
> 'includeantruntime' was not set, defaulting to
> build.sysclasspath=last; set to false for repeatable builds
>    [javac] Compiling 282 source files to
> /home/todd/git/hadoop-common/hdfs/build/classes
>    [javac] 
> /home/todd/git/hadoop-common/hdfs/src/java/org/apache/hadoop/fs/Hdfs.java:33:
> package org.apache.hadoop.conf does not exist
>    [javac] import org.apache.hadoop.conf.Configuration;
> 
> ... [lots more errors where o.a.h.* from common can't be found ...
> 
> and yet:
> 
> todd@todd-w510:~/git/hadoop-common/hdfs$ find . -name \*common\*23\*jar
> ./build/ivy/lib/hadoop-hdfs/common/hadoop-common-0.23.0-SNAPSHOT.jar
> ./build/ivy/lib/hadoop-hdfs/test/hadoop-common-0.23.0-SNAPSHOT-tests.jar
> 
> and then I run the exact build command again, I see:
> 
> ivy-resolve-common:
> [ivy:resolve] downloading
> https://repository.apache.org/content/repositories/snapshots/org/apache/hadoop/hadoop-common/0.23.0-SNAPSHOT/hadoop-common-0.23.0-20110815.225725-267.jar
> ...
> [ivy:resolve] 
> ...............................................................................................................................
> [ivy:resolve] 
> .........................................................................................
> (1667kB)
> [ivy:resolve] .. (0kB)
> [ivy:resolve]   [SUCCESSFUL ]
> org.apache.hadoop#hadoop-common;0.23.0-SNAPSHOT!hadoop-common.jar
> (2661ms)
> 
> and it builds fine.
> 
> 
> On Mon, Aug 15, 2011 at 4:08 PM, Giridharan Kesavan
> <gkesa...@hortonworks.com> wrote:
>> On Mon, Aug 15, 2011 at 3:52 PM, Eli Collins <e...@cloudera.com> wrote:
>>> On Mon, Aug 15, 2011 at 3:45 PM, Giridharan Kesavan
>>> <gkesa...@hortonworks.com> wrote:
>>>> Hi Eli,
>>>> 
>>>> your are right Im talking about the apache jenkins hdfs build failures;
>>>> 
>>>> Im pretty sure hdfs is picking the latest hadoop common jars. I
>>>> verified this with the apache repo as well.
>>> 
>>> How are you building? The method that its claiming doesn't exist
>>> definitely does.
>> 
>> target doesn't exist in the build.xml; its part of the fault injection
>> framework and is imported from "trunk/hdfs/src/test/aop/build/aop.xml"
>> 
>> you can see this import in the build.xml file
>>  <import file="${test.src.dir}/aop/build/aop.xml"/>
>> 
>>> 
>>> The following works on trunk so I think it's  an issue with how
>>> Jenkins is running it.
>>> 
>>> hadoop-common $ mvn clean
>>> hadoop-common $ mvn install -DskipTests
>>> hadoop-common $ pushd ../hdfs
>>> hdfs $ ant clean
>>> hdfs $ ant -Dresolvers=internal jar
>>> hdfs $ ant run-test-hdfs-fault-inject
>> 
>>  I think you should pass "-Dresolver=internal" to the
>> run-test-hdfs-fault-inject target as well
>> 
>> Thanks,
>> giri
>> 
>>> 
>>> Thanks,
>>> Eli
>>> 
>>>> 
>>>> https://repository.apache.org/content/groups/snapshots/org/apache/hadoop/hadoop-common/0.23.0-SNAPSHOT
>>>> 
>>>> hadoop-common-0.23.0-20110815.215733-266-tests.jar
>>>> 
>>>> hadoop-common-0.23.0-20110815.215733-266.jar
>>>> 
>>>> https://builds.apache.org/view/G-L/view/Hadoop/job/Hadoop-Hdfs-trunk-Commit/837/console
>>>> 
>>>> [ivy:resolve] .. (0kB)
>>>> 
>>>> [ivy:resolve]   [SUCCESSFUL ] org.apache.hadoop#avro;1.3.2!avro.jar 
>>>> (1011ms)
>>>> [ivy:resolve] downloading
>>>> https://repository.apache.org/content/repositories/snapshots/org/apache/hadoop/hadoop-common/0.23.0-SNAPSHOT/hadoop-common-0.23.0-20110815.215733-266.jar
>>>> ...
>>>> 
>>>> [ivy:resolve] 
>>>> ........................................................................................................................................................................................................................
>>>> (1667kB)
>>>> [ivy:resolve] .. (0kB)
>>>> [ivy:resolve]   [SUCCESSFUL ]
>>>> org.apache.hadoop#hadoop-common;0.23.0-SNAPSHOT!hadoop-common.jar
>>>> (1549ms)
>>>> 
>>>> ivy-retrieve-common:
>>>> [ivy:cachepath] DEPRECATED: 'ivy.conf.file' is deprecated, use
>>>> 'ivy.settings.file' instead
>>>> [ivy:cachepath] :: loading settings :: file =
>>>> /home/jenkins/jenkins-slave/workspace/Hadoop-Hdfs-trunk-Commit/trunk/ivy/ivysettings.xml
>>>> 
>>>> ivy-resolve-hdfs:
>>>> 
>>>> ivy-retrieve-hdfs:
>>>> 
>>>> ivy-resolve-test:
>>>> 
>>>> [ivy:resolve] downloading
>>>> https://repository.apache.org/content/repositories/snapshots/org/apache/hadoop/hadoop-common/0.23.0-SNAPSHOT/hadoop-common-0.23.0-20110815.215733-266-tests.jar
>>>> ...
>>>> [ivy:resolve] 
>>>> ........................................................................................................................
>>>> (876kB)
>>>> 
>>>> [ivy:resolve] .. (0kB)
>>>> [ivy:resolve]   [SUCCESSFUL ]
>>>> org.apache.hadoop#hadoop-common;0.23.0-SNAPSHOT!hadoop-common.jar(tests)
>>>> (875ms)
>>>> 
>>>> 
>>>> On Mon, Aug 15, 2011 at 3:33 PM, Eli Collins <e...@cloudera.com> wrote:
>>>>> 
>>>>> Hey Giri,
>>>>> 
>>>>> This looks like a similar issue to what was hitting the main Jenkins
>>>>> job, the Hdfs job isn't picking up the latest bits from common.
>>>>> 
>>>>> Thanks,
>>>>> Eli
>>>>> 
>>>>> On Mon, Aug 15, 2011 at 3:27 PM, Giridharan Kesavan
>>>>> <gkesa...@hortonworks.com> wrote:
>>>>>> Todd,
>>>>>> 
>>>>>> Could you please take a look at this ?
>>>>>> 
>>>>>> https://issues.apache.org/jira/browse/HDFS-2261
>>>>>> 
>>>>>> 
>>>>>> -Giri
>>>>>> On Mon, Aug 15, 2011 at 3:24 PM, Todd Lipcon <t...@cloudera.com> wrote:
>>>>>> 
>>>>>>> Seems like some of it is a build issue where it can't find ant.
>>>>>>> 
>>>>>>> The other is the following:
>>>>>>> https://issues.apache.org/jira/browse/HADOOP-7545
>>>>>>> Please review.
>>>>>>> 
>>>>>>> Thanks
>>>>>>> -Todd
>>>>>>> 
>>>>>>> On Mon, Aug 15, 2011 at 2:54 PM, Mahadev Konar <maha...@hortonworks.com>
>>>>>>> wrote:
>>>>>>>> Hi folks,
>>>>>>>>  Can anyone take a look at the hdfs builds? Seems to be failing:
>>>>>>>> 
>>>>>>>> https://builds.apache.org/job/Hadoop-Hdfs-trunk/
>>>>>>>> 
>>>>>>>> thanks
>>>>>>>> mahadev
>>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> --
>>>>>>> Todd Lipcon
>>>>>>> Software Engineer, Cloudera
>>>>>>> 
>>>>>> 
>>>> 
>>> 
>> 
> 
> 
> 
> -- 
> Todd Lipcon
> Software Engineer, Cloudera

Reply via email to