It worked fine when i updated /etc/hosts file (of all the slaves) and writing fully qualified domain name in the hadoop-site.xml.
It worked fine for sometime .. then started giving new error 09/06/23 22:21:49 INFO ipc.Client: Retrying connect to server: master/ 10.2.24.21:54310. Already tried 0 time(s). 09/06/23 22:21:50 INFO ipc.Client: Retrying connect to server: master/ 10.2.24.21:54310. Already tried 1 time(s). 09/06/23 22:21:51 INFO ipc.Client: Retrying connect to server: master/ 10.2.24.21:54310. Already tried 2 time(s). 09/06/23 22:21:52 INFO ipc.Client: Retrying connect to server: master/ 10.2.24.21:54310. Already tried 3 time(s). 09/06/23 22:21:53 INFO ipc.Client: Retrying connect to server: master/ 10.2.24.21:54310. Already tried 4 time(s). 09/06/23 22:21:54 INFO ipc.Client: Retrying connect to server: master/ 10.2.24.21:54310. Already tried 5 time(s). 09/06/23 22:21:55 INFO ipc.Client: Retrying connect to server: master/ 10.2.24.21:54310. Already tried 6 time(s). 09/06/23 22:21:56 INFO ipc.Client: Retrying connect to server: master/ 10.2.24.21:54310. Already tried 7 time(s). 09/06/23 22:21:57 INFO ipc.Client: Retrying connect to server: master/ 10.2.24.21:54310. Already tried 8 time(s). 09/06/23 22:21:58 INFO ipc.Client: Retrying connect to server: master/ 10.2.24.21:54310. Already tried 9 time(s). On Tue, Jun 23, 2009 at 8:33 PM, Raghu Angadi <rang...@yahoo-inc.com> wrote: > Raghu Angadi wrote: > >> >> This is at RPC client level and there is requirement for fully qualified >> > > I meant to say "there is NO requirement ..." > > hostname. May be "." at the end of "10.2.24.21" causing the problem? >> >> btw, in 0.21 even fs.default.name does not need to be fully qualified >> > > that fix is probably in 0.20 too. > > Raghu. > > > name.. anything that resolves to an ipaddress is fine (at least for >> common/FS and HDFS). >> >> Raghu. >> >> Matt Massie wrote: >> >>> fs.default.name in your hadoop-site.xml needs to be set to a >>> fully-qualified domain name (instead of an IP address) >>> >>> -Matt >>> >>> On Jun 23, 2009, at 6:42 AM, bharath vissapragada wrote: >>> >>> when i try to execute the command bin/start-dfs.sh , i get the >>>> following >>>> error . I have checked the hadoop-site.xml file on all the nodes , and >>>> they >>>> are fine .. >>>> can some-one help me out! >>>> >>>> 10.2.24.21: Exception in thread "main" java.net.UnknownHostException: >>>> unknown host: 10.2.24.21. >>>> 10.2.24.21: at >>>> org.apache.hadoop.ipc.Client$Connection.<init>(Client.java:195) >>>> 10.2.24.21: at >>>> org.apache.hadoop.ipc.Client.getConnection(Client.java:779) >>>> 10.2.24.21: at org.apache.hadoop.ipc.Client.call(Client.java:704) >>>> 10.2.24.21: at >>>> org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216) >>>> 10.2.24.21: at >>>> org.apache.hadoop.dfs.$Proxy4.getProtocolVersion(Unknown >>>> Source) >>>> 10.2.24.21: at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:319) >>>> 10.2.24.21: at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:306) >>>> 10.2.24.21: at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:343) >>>> 10.2.24.21: at org.apache.hadoop.ipc.RPC.waitForProxy(RPC.java:288) >>>> >>> >>> >> >> >