please correct your hdfs-site.xml contents: <configuration> <property> <name>dfs.name.dir</name> <value>/home/hadoop/tmp</value> </property>
<property> <name>dfs.data.dir</name> <value>/home/hadoop/data</value> </property> <property> <name>dfs.replication</name> <value>1</value> </property> </configuration> rest all things are fine one more thing:it is best practice that you edit and add details of your matsre and slave (which you can get from my blog ) in : /etc/hosts conf/master conf/slaves 2010/11/25 祝美祺 <henhaozhume...@gmail.com> > thanks for your reply,here is my configuration. > master: > > core-site.xml: > > <configuration> > <property> > <name>fs.default.name</name> > <value>hdfs://192.168.0.142:9000</value> > </property> > > </configuration> > > > hdfs-site.xml: > > <configuration> > <property> > <name>dfs.name.dir</name> > <value>/home/hadoop/tmp</value> > </property> > <property> > <property> > <name>dfs.data.dir</name> > <value>/home/hadoop/data</value> > </property> > <name>dfs.replication</name> > <value>1</value> > </property> > > </configuration> > > mapred-site.xml: > > <configuration> > <property> > <name>mapred.job.tracker</name> > <value>192.168.0.142:9200</value> > <description>jobtracker host port</description> > </property> > </configuration> > > > And they are the same on slaves.And any problem with these configuration.i > got only one slave on 192.168.0.177 > > 2010/11/25 rahul patodi <patodira...@gmail.com> > > also check your other file like /etc/hosts >> >> >> On Thu, Nov 25, 2010 at 10:36 AM, rahul patodi <patodira...@gmail.com>wrote: >> >>> i think you should check your configuration file in the conf folder and >>> add the required entry in >>> core-site.xml, mapred-site.xml and hdfs-site.xml >>> for pseudo distributed mode you can refer: >>> http://hadoop-tutorial.blogspot.com/2010/11/running-hadoop-in-pseudo-distributed.html >>> for distributed mode you can refer: >>> http://hadoop-tutorial.blogspot.com/2010/11/running-hadoop-in-distributed-mode.html >>> >>> if you are using cloudera you can refer: >>> http://cloudera-tutorial.blogspot.com/ >>> >>> if you have any problem please leave a comment >>> >>> >>> -- >>> -Thanks and Regards, >>> Rahul Patodi >>> Associate Software Engineer, >>> Impetus Infotech (India) Private Limited, >>> www.impetus.com >>> Mob:09907074413 >>> >>> On Thu, Nov 25, 2010 at 10:23 AM, 祝美祺 <henhaozhume...@gmail.com> wrote: >>> >>>> Hi , >>>> i am using 0.20.2.. when i try to start hadoop,the namenode >>>> ,datanode works will ,but i can't submit jobs.i look at the logs ,and find >>>> the error like: >>>> >>>> 2010-11-25 12:38:02,623 INFO org.apache.hadoop.ipc.Client: Retrying >>>> connect to server: localhost/127.0.0.1:9000. Already tried 0 time(s). >>>> 2010-11-25 12:38:03,626 INFO org.apache.hadoop.ipc.Client: Retrying >>>> connect to server: localhost/127.0.0.1:9000. Already tried 1 time(s). >>>> 2010-11-25 12:38:04,627 INFO org.apache.hadoop.ipc.Client: Retrying >>>> connect to server: localhost/127.0.0.1:9000. Already tried 2 time(s). >>>> 2010-11-25 12:38:05,629 INFO org.apache.hadoop.ipc.Client: Retrying >>>> connect to server: localhost/127.0.0.1:9000. Already tried 3 time(s). >>>> 2010-11-25 12:38:06,631 INFO org.apache.hadoop.ipc.Client: Retrying >>>> connect to server: localhost/127.0.0.1:9000. Already tried 4 time(s). >>>> 2010-11-25 12:38:07,636 INFO org.apache.hadoop.ipc.Client: Retrying >>>> connect to server: localhost/127.0.0.1:9000. Already tried 5 time(s). >>>> 2010-11-25 12:38:08,641 INFO org.apache.hadoop.ipc.Client: Retrying >>>> connect to server: localhost/127.0.0.1:9000. Already tried 6 time(s). >>>> 2010-11-25 12:38:09,644 INFO org.apache.hadoop.ipc.Client: Retrying >>>> connect to server: localhost/127.0.0.1:9000. Already tried 7 time(s). >>>> 2010-11-25 12:38:10,645 INFO org.apache.hadoop.ipc.Client: Retrying >>>> connect to server: localhost/127.0.0.1:9000. Already tried 8 time(s). >>>> 2010-11-25 12:38:11,649 INFO org.apache.hadoop.ipc.Client: Retrying >>>> connect to server: localhost/127.0.0.1:9000. Already tried 9 time(s). >>>> 2010-11-25 12:38:11,651 ERROR org.apache.hadoop.mapred.TaskTracker: >>>> Caught exception: java.net.ConnectException: Call to localhost/ >>>> 127.0.0.1:9000 failed on connection exception: >>>> java.net.ConnectException: Connection refused >>>> at org.apache.hadoop.ipc.Client.wrapException(Client.java:767) >>>> at org.apache.hadoop.ipc.Client.call(Client.java:743) >>>> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:220) >>>> at $Proxy5.getProtocolVersion(Unknown Source) >>>> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:359) >>>> at >>>> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:106) >>>> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:207) >>>> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:170) >>>> at >>>> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:82) >>>> at >>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1378) >>>> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66) >>>> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1390) >>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:196) >>>> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:175) >>>> at >>>> org.apache.hadoop.mapred.TaskTracker.offerService(TaskTracker.java:1033) >>>> at org.apache.hadoop.mapred.TaskTracker.run(TaskTracker.java:1720) >>>> at org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:2833) >>>> Caused by: java.net.ConnectException: Connection refused >>>> at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) >>>> at >>>> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:592) >>>> at >>>> org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206) >>>> at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:404) >>>> at >>>> org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:304) >>>> at >>>> org.apache.hadoop.ipc.Client$Connection.access$1700(Client.java:176) >>>> at org.apache.hadoop.ipc.Client.getConnection(Client.java:860) >>>> at org.apache.hadoop.ipc.Client.call(Client.java:720) >>>> ... 15 more >>>> >>>> Who can tell me how to solve this problem? >>>> >>>> >>> >>> >>> >> >> >> > > > -- > 祝美祺 > > 电话:010-62789307 > 清华大学计算机科学与技术系 > 清华大学伟清楼709 > 北京,100084 > > Zhu Mei qi > > Tel.: +86-10-62789307 > Room 709, WeiQing Building > Tsinghua University, Beijing 100084, China > -- -Thanks and Regards, Rahul Patodi Associate Software Engineer, Impetus Infotech (India) Private Limited, www.impetus.com Mob:09907074413