You need to set JAVA_HOME to your jdk directory (instead of jre). This is required by ant.
Zheng -----Original Message----- From: Sandy [mailto:[EMAIL PROTECTED] Sent: Wednesday, June 25, 2008 11:22 AM To: [email protected] Subject: Re: Compiling Word Count in C++ : Hadoop Pipes I'm not sure how this answers my question. Could you be more specific? I still am getting the above error when I type this commmand in. To summarize: With my current setup, this occurs: $ ant -Dcompile.c++=yes compile-c++-examples Unable to locate tools.jar. Expected to find it in /usr/java/jre1.6.0_06/lib/tools.jar Buildfile: build.xml init: [touch] Creating /tmp/null2044923713 [delete] Deleting: /tmp/null2044923713 [exec] svn: '.' is not a working copy [exec] svn: '.' is not a working copy check-c++-makefiles: create-c++-examples-pipes-makefile: [mkdir] Created dir: /home/sjm/Desktop/hadoop-0.16.4/build/c++-build/Linux-i386-32/examples/p ipes BUILD FAILED /home/sjm/Desktop/hadoop-0.16.4/build.xml:987: Execute failed: java.io.IOException: Cannot run program "/home/sjm/Desktop/hadoop-0.16.4/src/examples/pipes/configure" (in directory "/home/sjm/Desktop/hadoop-0.16.4/build/c++-build/Linux-i386-32/examples/ pipes"): java.io.IOException: error=13, Permission denied Total time: 1 second ----- If I copy the tools.jar file located in my jdk's lib folder, i get the error message I printed in the previous message. Could someone please tell me or suggest to me what I am doing wrong? Thanks, -SM On Wed, Jun 25, 2008 at 1:53 PM, lohit <[EMAIL PROTECTED]> wrote: > ant -Dcompile.c++=yes compile-c++-examples > I picked it up from build.xml > > Thanks, > Lohit > > ----- Original Message ---- > From: Sandy <[EMAIL PROTECTED]> > To: [email protected] > Sent: Wednesday, June 25, 2008 10:44:20 AM > Subject: Compiling Word Count in C++ : Hadoop Pipes > > Hi, > > I am currently trying to get Hadoop Pipes working. I am following the > instructions at the hadoop wiki, where it provides code for a C++ > implementation of Word Count (located here: > http://wiki.apache.org/hadoop/C++WordCount?highlight=%28C%2B%2B%29) > > I am having some trouble parsing the instructions. What should the file > containing the new word count program be called? "examples"? > > If I were to call the file "example" and type in the following: > $ ant -Dcompile.c++=yes example > Buildfile: build.xml > > BUILD FAILED > Target `example' does not exist in this project. > > Total time: 0 seconds > > > If I try and compile with "examples" as stated on the wiki, I get: > $ ant -Dcompile.c++=yes examples > Buildfile: build.xml > > clover.setup: > > clover.info: > [echo] > [echo] Clover not found. Code coverage reports disabled. > [echo] > > clover: > > init: > [touch] Creating /tmp/null810513231 > [delete] Deleting: /tmp/null810513231 > [exec] svn: '.' is not a working copy > [exec] svn: '.' is not a working copy > > record-parser: > > compile-rcc-compiler: > [javac] Compiling 29 source files to > /home/sjm/Desktop/hadoop-0.16.4/build/classes > > BUILD FAILED > /home/sjm/Desktop/hadoop-0.16.4/build.xml:241: Unable to find a javac > compiler; > com.sun.tools.javac.Main is not on the classpath. > Perhaps JAVA_HOME does not point to the JDK > > Total time: 1 second > > > > I am a bit puzzled by this. Originally I got the error that tools.jar was > not found, because it was looking for it under > /usr/java/jre1.6.0_06/lib/tools.jar . There is a tools.jar under > /usr/java/jdk1.6.0_06/lib/tools.jar. If I copy this file over to the jre > folder, that message goes away and its replaced with the above message. > > My hadoop-env.sh file looks something like: > # Set Hadoop-specific environment variables here. > > # The only required environment variable is JAVA_HOME. All others are > # optional. When running a distributed configuration it is best to > # set JAVA_HOME in this file, so that it is correctly defined on > # remote nodes. > > # The java implementation to use. Required. > # export JAVA_HOME=$JAVA_HOME > > > and my .bash_profile file has this line in it: > JAVA_HOME=/usr/java/jre1.6.0_06; export JAVA_HOME > export PATH > > > Furthermore, if I go to the command line and type in javac -version, I get: > $ javac -version > javac 1.6.0_06 > > > I also had no problem getting through the hadoop word count map reduce > tutorial in Java. It was able to find my java compiler fine. Could someone > please point me in the right direction? Also, since it is an sh file, > should > that export line in hadoop-env.sh really start with a hash sign? > > Thank you in advance for your assistance. > > -SM > >
