Ubuntu seems to install the protocol buffer library (protobuf-compiler) as part of the standard install. Can you run 'protoc --version' to figure out which version is being used. If you've installed it separately - you could play around with the path, remove the package installed by Ubuntu, etc to make sure protoc 2.4 is used.
- Sid On Fri, Jun 17, 2011 at 1:15 AM, Luke Lu <[email protected]> wrote: > MR-279 actually works fine with maven 3.0.3 (sans a few (IMO bogus) > warnings). You can leave out the "tar" target (which depends on the > "docs" target, which requires forrest 0.8) to unblock the progress, as > mvn-install would suffice for common and hdfs builds. > > On Thu, Jun 16, 2011 at 7:55 PM, Praveen Sripati > <[email protected]> wrote: > > Tom, > > > > I downgraded maven and also changed from open-jdk to sun-jdk and there is > > not any progress. I am using Ubuntu 11.04 and could not find > sun-java5-jdk > > in the Ubuntu repositories, so I installed sun-java6-jdk. > > > > praveensripati@praveensripati:~$ java -version > > java version "1.6.0_24" > > Java(TM) SE Runtime Environment (build 1.6.0_24-b07) > > > > praveensripati@praveensripati:~$ mvn -version > > Apache Maven 2.2.1 (r801777; 2009-08-07 00:46:01+0530) > > > > Thanks, > > Praveen > > > > > > On Friday 17 June 2011 02:04 AM, Thomas Graves wrote: > >> > >> I know at one time maven 3.x didn't work so I've been using maven 2.x. > >> > >> Well I've never tried using java6 for java5 home but I would think it > >> wouldn't work. I thought it was forrest that required java5. I would > >> suggest using java5. > >> > >> Tom > >> > >> > >> On 6/16/11 12:24 PM, "Praveen Sripati"<[email protected]> > wrote: > >> > >>> Tom, > >>> > >>>>> Note, it looks like your java5.home is pointing to java6? > >>> > >>> I have java6 on my laptop and pointed java5.home variable to java6. The > >>> hadoop doc says "Java 1.6.x - preferable from Sun". Is this the > problem? > >>> > >>>>> What version of protobufs are you using? > >>> > >>> I have protobuf 2.4.1. > >>> > >>>>> What about mvn version? > >>> > >>> Apache Maven 3.0.3 (r1075438; 2011-02-28 23:01:09+0530) > >>> > >>>>> So you had both common and hdfs built before doing mapreduce and > >>> > >>> common built before building hdfs? Or was common failing with the error > >>> you mention below? If you haven't already you might simply try > >>> veryclean on everything and go again in order. > >>> I tried common first and there were some errors related to fop, but the > >>> common jars were created, so I started with hdfs and it was successful. > >>> Then I started the yarn build which led to the > >>> java_generate_equals_and_hash error. > >>> > >>> Thanks, > >>> Praveen > >>> > >>> > >>> On Thursday 16 June 2011 09:54 PM, Thomas Graves wrote: > >>>> > >>>> Note, it looks like your java5.home is pointing to java6? > >>>> > >>>> I've never seen this particular error. The > java_generate_equals_and_hash > >>>> option seems to have been added in protobuf2.4.0. What version of > >>>> protobufs > >>>> are you using? The instructions say to use atleast 2.4.0a, I'm using > >>>> 2.4.1 > >>>> right now. > >>>> > >>>> You need to define the following (I use a build.properties file). > These > >>>> are > >>>> the version I'm currently using. All of these are just downloaded > from > >>>> the > >>>> corresponding website. Some links to those can be found here: > >>>> http://yahoo.github.com/hadoop-common/installing.html > >>>> > >>>> java5.home=/home/tgraves/hadoop/jdk1.5.0_22/ > >>>> forrest.home=/home/tgraves/hadoop/apache-forrest-0.8 > >>>> ant.home=/home/tgraves/hadoop/apache-ant-1.8.2 > >>>> xercescroot=/home/tgraves/hadoop/xerces-c-src_2_8_0 > >>>> eclipse.home=/home/tgraves/hadoop/eclipse > >>>> findbugs.home=/home/tgraves/hadoop/findbugs-1.3.9 > >>>> > >>>> I thought this was the same as for trunk but perhaps I'm mistaken. > >>>> > >>>> What about mvn version? > >>>> /home/y/libexec/maven/bin/mvn --version > >>>> Apache Maven 2.2.1 (r801777; 2009-08-06 19:16:01+0000) > >>>> > >>>> So you had both common and hdfs built before doing mapreduce and > common > >>>> built before building hdfs? Or was common failing with the error you > >>>> mention > >>>> below? If you haven't already you might simply try veryclean on > >>>> everything > >>>> and go again in order. > >>>> > >>>> Tom > >>>> > >>>> > >>>> On 6/16/11 8:10 AM, "Praveen Sripati"<[email protected]> > wrote: > >>>> > >>>>> Hi, > >>>>> > >>>>> The hdfs build was successful after including the -Dforrest.home > >>>>> property to the ant command. > >>>>> > >>>>> *********** > >>>>> > >>>>> When I started the mapreduce build to get the below error. > >>>>> > >>>>> mvn clean install assembly:assembly > >>>>> > >>>>> Downloaded: > >>>>> > >>>>> > http://repo1.maven.org/maven2/org/apache/commons/commons-exec/1.0.1/commons- > >>>>> ex > >>>>> ec-1.0.1.jar > >>>>> (49 KB at 24.4 KB/sec) > >>>>> yarn_protos.proto:4:8: Option "java_generate_equals_and_hash" > unknown. > >>>>> [INFO] > >>>>> > >>>>> [INFO] > >>>>> > >>>>> > ------------------------------------------------------------------------ > >>>>> [INFO] Skipping hadoop-mapreduce > >>>>> [INFO] This project has been banned from the build due to previous > >>>>> failures. > >>>>> [INFO] > >>>>> > >>>>> > ------------------------------------------------------------------------ > >>>>> [INFO] > >>>>> > >>>>> > ------------------------------------------------------------------------ > >>>>> [INFO] Reactor Summary: > >>>>> [INFO] > >>>>> [INFO] yarn-api .......................................... FAILURE > >>>>> [13:23.081s] > >>>>> [INFO] yarn-common ....................................... SKIPPED > >>>>> [INFO] yarn-server-common ................................ SKIPPED > >>>>> [INFO] yarn-server-nodemanager ........................... SKIPPED > >>>>> [INFO] yarn-server-resourcemanager ....................... SKIPPED > >>>>> [INFO] yarn-server-tests ................................. SKIPPED > >>>>> [INFO] yarn-server ....................................... SKIPPED > >>>>> [INFO] yarn .............................................. SKIPPED > >>>>> [INFO] hadoop-mapreduce-client-core ...................... SKIPPED > >>>>> [INFO] hadoop-mapreduce-client-common .................... SKIPPED > >>>>> [INFO] hadoop-mapreduce-client-shuffle ................... SKIPPED > >>>>> [INFO] hadoop-mapreduce-client-app ....................... SKIPPED > >>>>> [INFO] hadoop-mapreduce-client-hs ........................ SKIPPED > >>>>> [INFO] hadoop-mapreduce-client-jobclient ................. SKIPPED > >>>>> [INFO] hadoop-mapreduce-client ........................... SKIPPED > >>>>> [INFO] hadoop-mapreduce .................................. SKIPPED > >>>>> [INFO] > >>>>> > >>>>> > ------------------------------------------------------------------------ > >>>>> [INFO] BUILD FAILURE > >>>>> [INFO] > >>>>> > >>>>> > ------------------------------------------------------------------------ > >>>>> [INFO] Total time: 13:45.437s > >>>>> [INFO] Finished at: Thu Jun 16 18:30:48 IST 2011 > >>>>> [INFO] Final Memory: 6M/15M > >>>>> [INFO] > >>>>> > >>>>> > ------------------------------------------------------------------------ > >>>>> [ERROR] Failed to execute goal > >>>>> org.codehaus.mojo:exec-maven-plugin:1.2:exec (generate-sources) on > >>>>> project yarn-api: Command execution failed. Process exited with an > >>>>> error: 1(Exit value: 1) -> [Help 1] > >>>>> [ERROR] > >>>>> [ERROR] To see the full stack trace of the errors, re-run Maven with > >>>>> the > >>>>> -e switch. > >>>>> [ERROR] Re-run Maven using the -X switch to enable full debug > logging. > >>>>> [ERROR] > >>>>> [ERROR] For more information about the errors and possible solutions, > >>>>> please read the following articles: > >>>>> [ERROR] [Help 1] > >>>>> > http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException > >>>>> > >>>>> *********** > >>>>> > >>>>> I started building the commons and had to include -Djava5.home and > >>>>> -Dforrest.home properties in the ant command. > >>>>> > >>>>> ant -Djava5.home=/usr/lib/jvm/java-6-openjdk > >>>>> -Dforrest.home=/home/praveensripati/Installations/apache-forrest-0.9 > >>>>> veryclean mvn-install tar > >>>>> > >>>>> And then I get the below error and the build hangs , but I see 4 jars > >>>>> in > >>>>> the build folder including hadoop-common-0.22.0-SNAPSHOT.jar. > >>>>> > >>>>> [exec] Cocoon will report the status of each document: > >>>>> [exec] - in column 1: *=okay X=brokenLink ^=pageSkipped (see > >>>>> FAQ). > >>>>> [exec] > >>>>> [exec] > >>>>> > >>>>> > ------------------------------------------------------------------------ > >>>>> [exec] cocoon 2.1.12-dev > >>>>> [exec] Copyright (c) 1999-2007 Apache Software Foundation. All > >>>>> rights reserved. > >>>>> [exec] > >>>>> > >>>>> > ------------------------------------------------------------------------ > >>>>> [exec] > >>>>> [exec] > >>>>> [exec] * [1/29] [29/29] 6.547s 9.4Kb linkmap.html > >>>>> [exec] * [2/29] [1/28] 1.851s 22.3Kb hdfs_shell.html > >>>>> [exec] * [4/28] [1/28] 1.156s 21.1Kb distcp.html > >>>>> [exec] * [5/27] [0/0] 0.306s 0b distcp.pdf > >>>>> [exec] Exception in thread "main" > >>>>> java.lang.NoClassDefFoundError: > >>>>> org/apache//messaging/MessageHandler > >>>>> [exec] at > >>>>> > >>>>> > org.apache.cocoon.serialization.FOPSerializer.configure(FOPSerializer.java:1 > >>>>> 22 > >>>>> ) > >>>>> > >>>>> > >>>>> I included the following in the common/ivy.xml and the > >>>>> ./common/build/ivy/lib/Hadoop-Common/common/fop-0.93.jar file is > there. > >>>>> > >>>>> <dependency org="org.apache.xmlgraphics" > >>>>> name="fop" > >>>>> rev="${fop.version}" > >>>>> conf="common->default"/> > >>>>> > >>>>> and the following in the common/ivy/libraries.properties and still > get > >>>>> the same error. > >>>>> > >>>>> fop.version=0.93 > >>>>> > >>>>> Thanks, > >>>>> Praveen > >>>>> > >>>>> On Thursday 16 June 2011 07:55 AM, Luke Lu wrote: > >>>>>> > >>>>>> On Wed, Jun 15, 2011 at 6:45 PM, Praveen Sripati > >>>>>> <[email protected]> wrote: > >>>>>>> > >>>>>>> Do I need the avro-maven-plugin? When I ran the below command got > the > >>>>>>> error that the pom file was not found. Where do I get the jar and > the > >>>>>>> pom files for the avro-maven-plugin? I was able to get the source > >>>>>>> code > >>>>>>> for them, but not the binaries. > >>>>>>> > >>>>>>> mvn install:install-file > >>>>>>> -Dfile=./avro-maven-plugin/avro-maven-plugin-1.4.0-SNAPSHOT.jar > >>>>>>> -DpomFile=./avro-maven-plugin/avro-maven-plugin-1.4.0-SNAPSHOT.pom > >>>>>> > >>>>>> No, you no longer need to install avro-maven-plugin manually. It's > >>>>>> automatically installed via maven, as we switched to avro 1.5.1. > >>>>>> We'll fix the instruction. > >>>>>> > >>>>>> __Luke > > >
