Hello Subroto , There are multiple ways to install and set the environment variables for 2.x series.
Download the latest tar in your computer for Hadoop 2.0.x and unzip it to some directory lets say HADOOP_PREFIX Export the following environment variables in your computer export HADOOP_PREFIX="/home/hadoop/software/hadoop-2.0.0-alpha" export PATH=$PATH:$HADOOP_PREFIX/bin export PATH=$PATH:$HADOOP_PREFIX/sbin export HADOOP_MAPRED_HOME=${HADOOP_PREFIX} export HADOOP_COMMON_HOME=${HADOOP_PREFIX} export HADOOP_HDFS_HOME=${HADOOP_PREFIX} export YARN_HOME=${HADOOP_PREFIX} A detailed discussion for this is present at Jira for Hadoop RPMs why this project structure etc. Few days back i wrote tutorial on how to install 2.x series<http://jugnu-life.blogspot.in/2012/05/hadoop-20-install-tutorial-023x.html>, you can also have a look at that. Regards, Jagat Singh On Tue, Jun 5, 2012 at 7:12 PM, Subroto <ssan...@datameer.com> wrote: > Hi > > Following is the hadoop directory structure after extracting the tar ball. > I would like to know where and to which folder I need to set the > HADOOP_MAPRED_HOME, HADOOP_HDFS_HOME,HADOOP_COMMON_HOME,YARN_HOME etc so > that this cluster can be accessed within the cluster and from outside as > well. > > /usr/local/hadoop > bin > etc > hadoop > include > example > lib > libexec > sbin > share > doc > hadoop > common > hdfs > httpfs > mapreduce > tools > src > Cheers, > Subroto Sanyal