I have two nodes as Hadoop test. When I set fs.default.name to hdfs://hostname:54310/ in core-site.xml and mapred.job.tracker to hdfs://hostname:54311 in mapred-site.xml, I received the following error information while I started it by start-all.sh.
org.apache.hadoop.ipc.RemoteException: java.io.IOException: File /home/hadoop/tmp/mapred/system/jobtracker.info could only be replicated to 0 nodes, instead of 1 at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1271) at org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422) at sun.reflect.GeneratedMethodAccessor12.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) ... Then I had to change hdfs://hostname:54310/ to hdfs://ipAddress:54310/ and hdfs://hostname:54311 to hdfs://ipAddress:54311, it's ok while I started it by start-all.sh. However, when I run wordcount example, I got the following error message. java.lang.IllegalArgumentException: Wrong FS: hdfs://ipAddress:54310/home/hadoop/tmp/mapred/system/job_201101041628_0005/job.xml, expected: hdfs://hostname:54310 at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:310) at org.apache.hadoop.hdfs.DistributedFileSystem.checkPath(DistributedFileSystem.java:99) at org.apache.hadoop.hdfs.DistributedFileSystem.getPathName(DistributedFileSystem.java:155) at org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:453) at org.apache.hadoop.mapred.TaskTracker.localizeJob(TaskTracker.java:745) at org.apache.hadoop.mapred.TaskTracker.startNewTask(TaskTracker.java:1664) at org.apache.hadoop.mapred.TaskTracker.access$1200(TaskTracker.java:97) at org.apache.hadoop.mapred.TaskTracker$TaskLauncher.run(TaskTracker.java:1629) From message above, it seems the hdfs://hostname:port is not suitable for example run? What should I do ? Note: ipAddress means ip address I used, hostname means host name I used