thanks for your reply,here is my configuration. master: core-site.xml:
<configuration> <property> <name>fs.default.name</name> <value>hdfs://192.168.0.142:9000</value> </property> </configuration> hdfs-site.xml: <configuration> <property> <name>dfs.name.dir</name> <value>/home/hadoop/tmp</value> </property> <property> <property> <name>dfs.data.dir</name> <value>/home/hadoop/data</value> </property> <name>dfs.replication</name> <value>1</value> </property> </configuration> mapred-site.xml: <configuration> <property> <name>mapred.job.tracker</name> <value>192.168.0.142:9200</value> <description>jobtracker host port</description> </property> </configuration> And they are the same on slaves.And any problem with these configuration.i got only one slave on 192.168.0.177 2010/11/25 rahul patodi <patodira...@gmail.com> > also check your other file like /etc/hosts > > > On Thu, Nov 25, 2010 at 10:36 AM, rahul patodi <patodira...@gmail.com>wrote: > >> i think you should check your configuration file in the conf folder and >> add the required entry in >> core-site.xml, mapred-site.xml and hdfs-site.xml >> for pseudo distributed mode you can refer: >> http://hadoop-tutorial.blogspot.com/2010/11/running-hadoop-in-pseudo-distributed.html >> for distributed mode you can refer: >> http://hadoop-tutorial.blogspot.com/2010/11/running-hadoop-in-distributed-mode.html >> >> if you are using cloudera you can refer: >> http://cloudera-tutorial.blogspot.com/ >> >> if you have any problem please leave a comment >> >> >> -- >> -Thanks and Regards, >> Rahul Patodi >> Associate Software Engineer, >> Impetus Infotech (India) Private Limited, >> www.impetus.com >> Mob:09907074413 >> >> On Thu, Nov 25, 2010 at 10:23 AM, 祝美祺 <henhaozhume...@gmail.com> wrote: >> >>> Hi , >>> i am using 0.20.2.. when i try to start hadoop,the namenode >>> ,datanode works will ,but i can't submit jobs.i look at the logs ,and find >>> the error like: >>> >>> 2010-11-25 12:38:02,623 INFO org.apache.hadoop.ipc.Client: Retrying >>> connect to server: localhost/127.0.0.1:9000. Already tried 0 time(s). >>> 2010-11-25 12:38:03,626 INFO org.apache.hadoop.ipc.Client: Retrying >>> connect to server: localhost/127.0.0.1:9000. Already tried 1 time(s). >>> 2010-11-25 12:38:04,627 INFO org.apache.hadoop.ipc.Client: Retrying >>> connect to server: localhost/127.0.0.1:9000. Already tried 2 time(s). >>> 2010-11-25 12:38:05,629 INFO org.apache.hadoop.ipc.Client: Retrying >>> connect to server: localhost/127.0.0.1:9000. Already tried 3 time(s). >>> 2010-11-25 12:38:06,631 INFO org.apache.hadoop.ipc.Client: Retrying >>> connect to server: localhost/127.0.0.1:9000. Already tried 4 time(s). >>> 2010-11-25 12:38:07,636 INFO org.apache.hadoop.ipc.Client: Retrying >>> connect to server: localhost/127.0.0.1:9000. Already tried 5 time(s). >>> 2010-11-25 12:38:08,641 INFO org.apache.hadoop.ipc.Client: Retrying >>> connect to server: localhost/127.0.0.1:9000. Already tried 6 time(s). >>> 2010-11-25 12:38:09,644 INFO org.apache.hadoop.ipc.Client: Retrying >>> connect to server: localhost/127.0.0.1:9000. Already tried 7 time(s). >>> 2010-11-25 12:38:10,645 INFO org.apache.hadoop.ipc.Client: Retrying >>> connect to server: localhost/127.0.0.1:9000. Already tried 8 time(s). >>> 2010-11-25 12:38:11,649 INFO org.apache.hadoop.ipc.Client: Retrying >>> connect to server: localhost/127.0.0.1:9000. Already tried 9 time(s). >>> 2010-11-25 12:38:11,651 ERROR org.apache.hadoop.mapred.TaskTracker: >>> Caught exception: java.net.ConnectException: Call to localhost/ >>> 127.0.0.1:9000 failed on connection exception: >>> java.net.ConnectException: Connection refused >>> at org.apache.hadoop.ipc.Client.wrapException(Client.java:767) >>> at org.apache.hadoop.ipc.Client.call(Client.java:743) >>> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) >>> at $Proxy5.getProtocolVersion(Unknown Source) >>> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359) >>> at org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:106) >>> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:207) >>> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:170) >>> at >>> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:82) >>> at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378) >>> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66) >>> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390) >>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196) >>> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:175) >>> at >>> org.apache.hadoop.mapred.TaskTracker.offerService(TaskTracker.java:1033) >>> at org.apache.hadoop.mapred.TaskTracker.run(TaskTracker.java:1720) >>> at org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:2833) >>> Caused by: java.net.ConnectException: Connection refused >>> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) >>> at >>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:592) >>> at >>> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206) >>> at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:404) >>> at >>> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:304) >>> at org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:176) >>> at org.apache.hadoop.ipc.Client.getConnection(Client.java:860) >>> at org.apache.hadoop.ipc.Client.call(Client.java:720) >>> ... 15 more >>> >>> Who can tell me how to solve this problem? >>> >>> >> >> >> > > > -- 祝美祺 电话:010-62789307 清华大学计算机科学与技术系 清华大学伟清楼709 北京,100084 Zhu Mei qi Tel.: +86-10-62789307 Room 709, WeiQing Building Tsinghua University, Beijing 100084, China