commenting the line 127.0.0.1 in /etc/hosts is not working. if i format the namenode then automatically this line is added. any other solution?
On 16 October 2011 19:13, praveenesh kumar <[email protected]> wrote: > try commenting 127.0.0.1 localhost line in your /etc/hosts and then restart > the cluster and then try again. > > Thanks, > Praveenesh > > On Sun, Oct 16, 2011 at 2:00 PM, Humayun gmail <[email protected] > >wrote: > > > we are using hadoop on virtual box. when it is a single node then it > works > > fine for big dataset larger than the default block size. but in case of > > multinode cluster (2 nodes) we are facing some problems. > > Like when the input dataset is smaller than the default block size(64 MB) > > then it works fine. but when the input dataset is larger than the default > > block size then it shows ‘too much fetch failure’ in reduce state. > > here is the output link > > http://paste.ubuntu.com/707517/ > > > > From the above comments , there are many users who faced this problem. > > different users suggested to modify the /etc/hosts file in different > manner > > to fix the problem. but there is no ultimate solution.we need the actual > > solution thats why we are writing here. > > > > this is our /etc/hosts file > > 192.168.60.147 humayun # Added by NetworkManager > > 127.0.0.1 localhost.localdomain localhost > > ::1 humayun localhost6.localdomain6 localhost6 > > 127.0.1.1 humayun > > > > # The following lines are desirable for IPv6 capable hosts > > ::1 localhost ip6-localhost ip6-loopback > > fe00::0 ip6-localnet > > ff00::0 ip6-mcastprefix > > ff02::1 ip6-allnodes > > ff02::2 ip6-allrouters > > ff02::3 ip6-allhosts > > > > 192.168.60.1 master > > 192.168.60.2 slave > > >
