Thanks Jagat….
The tutorial is really nice ….

Cheers,
Subroto Sanyal

On Jun 6, 2012, at 9:47 AM, Jagat wrote:

> Hello Subroto , 
> 
> There are multiple ways to install and set the environment variables for 2.x 
> series.
> Download the latest tar in your computer for Hadoop 2.0.x and unzip it to 
> some directory lets say HADOOP_PREFIX
> 
> Export the following environment variables in your computer
> 
> export HADOOP_PREFIX="/home/hadoop/software/hadoop-2.0.0-alpha"
> export PATH=$PATH:$HADOOP_PREFIX/bin
> export PATH=$PATH:$HADOOP_PREFIX/sbin
> 
> export HADOOP_MAPRED_HOME=${HADOOP_PREFIX}
> export HADOOP_COMMON_HOME=${HADOOP_PREFIX}
> export HADOOP_HDFS_HOME=${HADOOP_PREFIX}
> export YARN_HOME=${HADOOP_PREFIX}
> 
> A detailed discussion for this is present at Jira for Hadoop RPMs why this 
> project structure etc.
> 
> Few days back i wrote tutorial on how to install 2.x series , you can also 
> have a look at that.
> 
> Regards,
> 
> Jagat Singh
> 
> On Tue, Jun 5, 2012 at 7:12 PM, Subroto <ssan...@datameer.com> wrote:
> Hi
> 
> Following is the hadoop directory structure after extracting the tar ball. I 
> would like to know where and to which folder I need to set the 
> HADOOP_MAPRED_HOME, HADOOP_HDFS_HOME,HADOOP_COMMON_HOME,YARN_HOME etc so that 
> this cluster can be accessed within the cluster and from outside as well.
> 
> /usr/local/hadoop
>                bin
>                etc
>                        hadoop
>                include
>                example
>                lib
>                libexec
>                sbin
>                share
>                        doc
>                        hadoop
>                                common
>                                hdfs
>                                httpfs
>                                mapreduce
>                                tools
>                src
> Cheers,
> Subroto Sanyal
> 

Reply via email to