Thank you very much! I have fixed that problem. There is something wrong with the host mapping... the default mapping is 127.0.0.1, not the exact ip address.
2010/10/16 Jander g <[email protected]> > Hi, > Didn't you format the HDFS ? or maybe the HDFS was broken. > > On Sat, Oct 16, 2010 at 9:01 PM, 冯超 <[email protected]> wrote: > > > Dear All, > > I met a problem that when I start the hadoop ,I could not find the > slave > > nodes , my OS system is CentOS, when run the mapreduce program, it alarms > > me > > that: > > > > > > 10/10/16 20:32:41 INFO ipc.Client: Retrying connect to server: zawc1/ > > 192.168.1.101:9000. Already tried 0 time(s). > > 10/10/16 20:32:42 INFO ipc.Client: Retrying connect to server: zawc1/ > > 192.168.1.101:9000. Already tried 1 time(s). > > 10/10/16 20:32:43 INFO ipc.Client: Retrying connect to server: zawc1/ > > 192.168.1.101:9000. Already tried 2 time(s). > > 10/10/16 20:32:44 INFO ipc.Client: Retrying connect to server: zawc1/ > > 192.168.1.101:9000. Already tried 3 time(s). > > 10/10/16 20:32:45 INFO ipc.Client: Retrying connect to server: zawc1/ > > 192.168.1.101:9000. Already tried 4 time(s). > > 10/10/16 20:32:46 INFO ipc.Client: Retrying connect to server: zawc1/ > > 192.168.1.101:9000. Already tried 5 time(s). > > 10/10/16 20:32:47 INFO ipc.Client: Retrying connect to server: zawc1/ > > 192.168.1.101:9000. Already tried 6 time(s). > > 10/10/16 20:32:48 INFO ipc.Client: Retrying connect to server: zawc1/ > > 192.168.1.101:9000. Already tried 7 time(s). > > 10/10/16 20:32:49 INFO ipc.Client: Retrying connect to server: zawc1/ > > 192.168.1.101:9000. Already tried 8 time(s). > > 10/10/16 20:32:50 INFO ipc.Client: Retrying connect to server: zawc1/ > > 192.168.1.101:9000. Already tried 9 time(s). > > java.net.ConnectException: Call to zawc1/192.168.1.101:9000 failed on > > connection exception: java.net.ConnectException: Connection refused > > at org.apache.hadoop.ipc.Client.wrapException(Client.java:724) > > at org.apache.hadoop.ipc.Client.call(Client.java:700) > > at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:216) > > at $Proxy0.getProtocolVersion(Unknown Source) > > at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:348) > > at > > org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:104) > > at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:176) > > at > > > > > org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:75) > > at > > org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1367) > > at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:56) > > at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1379) > > at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:215) > > at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:120) > > at > > org.apache.hadoop.examples.PiEstimator.launch(PiEstimator.java:187) > > at > org.apache.hadoop.examples.PiEstimator.run(PiEstimator.java:245) > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > > at > org.apache.hadoop.examples.PiEstimator.main(PiEstimator.java:252) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at > > > > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > > at > > > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > > at java.lang.reflect.Method.invoke(Method.java:597) > > at > > > > > org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.java:68) > > at > > org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:141) > > at > > org.apache.hadoop.examples.ExampleDriver.main(ExampleDriver.java:61) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at > > > > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > > at > > > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > > at java.lang.reflect.Method.invoke(Method.java:597) > > at org.apache.hadoop.util.RunJar.main(RunJar.java:165) > > at org.apache.hadoop.mapred.JobShell.run(JobShell.java:54) > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) > > at org.apache.hadoop.mapred.JobShell.main(JobShell.java:68) > > Caused by: java.net.ConnectException: Connection refused > > at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > > at > > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:574) > > at sun.nio.ch.SocketAdaptor.connect(SocketAdaptor.java:100) > > at > > org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:300) > > at > > org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:177) > > at org.apache.hadoop.ipc.Client.getConnection(Client.java:801) > > at org.apache.hadoop.ipc.Client.call(Client.java:686) > > ... 31 more > > > > Who can tell me what is the reason? Thanks for any help... > > > > > > -- > Thanks, > Jander >
