I am under the impression that it already is. As I posted in my original e-mail, here are the declarations in hadoop-env.sh and my .bash_profile
My hadoop-env.sh file looks something like: # Set Hadoop-specific environment variables here. # The only required environment variable is JAVA_HOME. All others are # optional. When running a distributed configuration it is best to # set JAVA_HOME in this file, so that it is correctly defined on # remote nodes. # The java implementation to use. Required. # export JAVA_HOME=$JAVA_HOME and my .bash_profile file has this line in it: JAVA_HOME=/usr/java/jre1.6.0_06; export JAVA_HOME export PATH Is there a different way I'm supposed to set the JAVA_HOME environment variable? Much thanks, -SM On Wed, Jun 25, 2008 at 3:22 PM, Zheng Shao <[EMAIL PROTECTED]> wrote: > You need to set JAVA_HOME to your jdk directory (instead of jre). > This is required by ant. > > Zheng > -----Original Message----- > From: Sandy [mailto:[EMAIL PROTECTED] > Sent: Wednesday, June 25, 2008 11:22 AM > To: [email protected] > Subject: Re: Compiling Word Count in C++ : Hadoop Pipes > > I'm not sure how this answers my question. Could you be more specific? I > still am getting the above error when I type this commmand in. To > summarize: > > With my current setup, this occurs: > $ ant -Dcompile.c++=yes compile-c++-examples > Unable to locate tools.jar. Expected to find it in > /usr/java/jre1.6.0_06/lib/tools.jar > Buildfile: build.xml > > init: > [touch] Creating /tmp/null2044923713 > [delete] Deleting: /tmp/null2044923713 > [exec] svn: '.' is not a working copy > [exec] svn: '.' is not a working copy > > check-c++-makefiles: > > create-c++-examples-pipes-makefile: > [mkdir] Created dir: > /home/sjm/Desktop/hadoop-0.16.4/build/c++-build/Linux-i386-32/examples/p > ipes > > BUILD FAILED > /home/sjm/Desktop/hadoop-0.16.4/build.xml:987: Execute failed: > java.io.IOException: Cannot run program > "/home/sjm/Desktop/hadoop-0.16.4/src/examples/pipes/configure" (in > directory > "/home/sjm/Desktop/hadoop-0.16.4/build/c++-build/Linux-i386-32/examples/ > pipes"): > java.io.IOException: error=13, Permission denied > > Total time: 1 second > > ----- > > If I copy the tools.jar file located in my jdk's lib folder, i get the > error > message I printed in the previous message. > > Could someone please tell me or suggest to me what I am doing wrong? > > Thanks, > > -SM > > On Wed, Jun 25, 2008 at 1:53 PM, lohit <[EMAIL PROTECTED]> wrote: > > > ant -Dcompile.c++=yes compile-c++-examples > > I picked it up from build.xml > > > > Thanks, > > Lohit > > > > ----- Original Message ---- > > From: Sandy <[EMAIL PROTECTED]> > > To: [email protected] > > Sent: Wednesday, June 25, 2008 10:44:20 AM > > Subject: Compiling Word Count in C++ : Hadoop Pipes > > > > Hi, > > > > I am currently trying to get Hadoop Pipes working. I am following the > > instructions at the hadoop wiki, where it provides code for a C++ > > implementation of Word Count (located here: > > http://wiki.apache.org/hadoop/C++WordCount?highlight=%28C%2B%2B%29) > > > > I am having some trouble parsing the instructions. What should the > file > > containing the new word count program be called? "examples"? > > > > If I were to call the file "example" and type in the following: > > $ ant -Dcompile.c++=yes example > > Buildfile: build.xml > > > > BUILD FAILED > > Target `example' does not exist in this project. > > > > Total time: 0 seconds > > > > > > If I try and compile with "examples" as stated on the wiki, I get: > > $ ant -Dcompile.c++=yes examples > > Buildfile: build.xml > > > > clover.setup: > > > > clover.info: > > [echo] > > [echo] Clover not found. Code coverage reports disabled. > > [echo] > > > > clover: > > > > init: > > [touch] Creating /tmp/null810513231 > > [delete] Deleting: /tmp/null810513231 > > [exec] svn: '.' is not a working copy > > [exec] svn: '.' is not a working copy > > > > record-parser: > > > > compile-rcc-compiler: > > [javac] Compiling 29 source files to > > /home/sjm/Desktop/hadoop-0.16.4/build/classes > > > > BUILD FAILED > > /home/sjm/Desktop/hadoop-0.16.4/build.xml:241: Unable to find a javac > > compiler; > > com.sun.tools.javac.Main is not on the classpath. > > Perhaps JAVA_HOME does not point to the JDK > > > > Total time: 1 second > > > > > > > > I am a bit puzzled by this. Originally I got the error that tools.jar > was > > not found, because it was looking for it under > > /usr/java/jre1.6.0_06/lib/tools.jar . There is a tools.jar under > > /usr/java/jdk1.6.0_06/lib/tools.jar. If I copy this file over to the > jre > > folder, that message goes away and its replaced with the above > message. > > > > My hadoop-env.sh file looks something like: > > # Set Hadoop-specific environment variables here. > > > > # The only required environment variable is JAVA_HOME. All others are > > # optional. When running a distributed configuration it is best to > > # set JAVA_HOME in this file, so that it is correctly defined on > > # remote nodes. > > > > # The java implementation to use. Required. > > # export JAVA_HOME=$JAVA_HOME > > > > > > and my .bash_profile file has this line in it: > > JAVA_HOME=/usr/java/jre1.6.0_06; export JAVA_HOME > > export PATH > > > > > > Furthermore, if I go to the command line and type in javac -version, I > get: > > $ javac -version > > javac 1.6.0_06 > > > > > > I also had no problem getting through the hadoop word count map reduce > > tutorial in Java. It was able to find my java compiler fine. Could > someone > > please point me in the right direction? Also, since it is an sh file, > > should > > that export line in hadoop-env.sh really start with a hash sign? > > > > Thank you in advance for your assistance. > > > > -SM > > > > >
