Ok, what about the issue regarding the users?  Do all the machines need to
be under the same user?

On Wed, Apr 23, 2008 at 12:43 PM, Harish Mallipeddi <
[EMAIL PROTECTED]> wrote:

> On Wed, Apr 23, 2008 at 3:03 PM, Sridhar Raman <[EMAIL PROTECTED]>
> wrote:
>
> > After trying out Hadoop in a single machine, I decided to run a
> MapReduce
> > across multiple machines.  This is the approach I followed:
> > 1 Master
> > 1 Slave
> >
> > (A doubt here:  Can my Master also be used to execute the Map/Reduce
> > functions?)
> >
>
> If you add the master node to the list of slaves (conf/slaves), then the
> master node run will also run a TaskTracker.
>
>
> >
> > To do this, I set up the masters and slaves files in the conf directory.
> > Following the instructions in this page -
> > http://hadoop.apache.org/core/docs/current/cluster_setup.html, I had set
> > up
> > sshd in both the machines, and was able to ssh from one to the other.
> >
> > I tried to run bin/start-dfs.sh.  Unfortunately, this asked for a
> password
> > for [EMAIL PROTECTED], while in slave, there was only user2.  While in 
> > master,
> > user1 was the logged on user.  How do I resolve this?  Should the user
> > accounts be present in all the machines?  Or can I specify this
> somewhere?
> >
>
>
>
> --
> Harish Mallipeddi
> circos.com : poundbang.in/blog/
>

Reply via email to