Yes it is required -- in psuedodistributed node the jobtracker is not necessarily aware that the task trackers / data nodes are on the same machine, and will thus attempt to ssh into them when starting the respective deamons etc (i.e. start-all.sh)
On Thu, May 16, 2013 at 11:21 AM, kishore alajangi < [email protected]> wrote: > When you start the hadoop procecess, each process will ask the password to > start, to overcome this we will configure SSH if you use single node or > multiple nodes for each process, if you can enter the password for each > process Its not a mandatory even if you use multiple systems. > > Thanks, > Kishore. > > > On Thu, May 16, 2013 at 8:24 PM, Raj Hadoop <[email protected]> wrote: > >> Hi, >> >> I have a dedicated user on Linux server for hadoop. I am installing it in >> psedo distributed mode on this box. I want to test my programs on this >> machine. But i see that in installation steps - they were mentioned that >> SSH needs to be configured. If it is single node, I dont require it >> ...right? Please advise. >> >> I was looking at this site >> >> http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/ >> >> It menionted like this - >> " >> Hadoop requires SSH access to manage its nodes, i.e. remote machines plus >> your local machine if you want to use Hadoop on it (which is what we want >> to do in this short tutorial). For our single-node setup of Hadoop, we >> therefore need to configure SSH access to localhost for the hduser user >> we created in the previous section. >> " >> >> Thanks, >> Raj >> >> > > -- Jay Vyas http://jayunit100.blogspot.com
