Hi, I have done a 2-node cluster setup by building hadoop-2.2-src. I am able to do the pseudo-distributed mode without any problem. But the cluster mode is creating problems. Node 1 (named 'compute-0-119.local') runs a Namenode, datanode, resourcemanage and a nodemanager Node 2 (named 'compute-0-118.local') runs a datanode and a nodemanager. I am able to bring up the daemons and the namenode, resourcemanager on Node 1 are able to connect to the daemons on Node 2. But on running a job I get the following Exception.
*14/06/20 22:24:41 ERROR security.UserGroupInformation: PriviledgedActionException as:srkandul (auth:SIMPLE) cause:java.net.ConnectException: Call From compute-0-119.local/10.1.255.13 <http://10.1.255.13> to localhost:8020 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused <http://wiki.apache.org/hadoop/ConnectionRefused>java.net.ConnectException: Call From compute-0-119.local/10.1.255.13 <http://10.1.255.13> to localhost:8020 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused <http://wiki.apache.org/hadoop/ConnectionRefused> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) at java.lang.reflect.Constructor.newInstance(Constructor.java:513) at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:783) at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:730) at org.apache.hadoop.ipc.Client.call(Client.java:1351) at org.apache.hadoop.ipc.Client.call(Client.java:1300) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206) * Following are the configurations I have set on both the machines. *core-site.xml* *<configuration><property> <name>fs.defaultFS</name> <value>hdfs://compute-0-119.local:8020</value> </property></configuration>* *mapred-site.xml* *<configuration> <property> <name>mapreduce.framework.name <http://mapreduce.framework.name></name> <value>yarn</value> </property> <property> <name>mapreduce.tasktracker.reduce.tasks.maximum</name> <value>8</value> </property></configuration>* *yarn-site.xml* *<configuration> <!-- Site specific YARN configuration properties --> <property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> <property> <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> <value>org.apache.hadoop.mapred.ShuffleHandler</value> </property> <property><name>yarn.resourcemanager.scheduler.address</name><value>compute-0-119.local:8030</value></property><property><name>yarn.resourcemanager.resource-tracker.address</name><value>compute-0-119.local:8031</value></property><property><name>yarn.resourcemanager.address</name><value>compute-0-119.local:8032</value></property><property><name>yarn.resourcemanager.admin.address</name><value>compute-0-119.local:8033</value></property><property><name>yarn.resourcemanager.webapp.address</name><value>compute-0-119.local:8088</value></property></configuration>* It looks like the client is still trying to connect to localhost:8020. Please help me with this. I am totally out of options here. I have googled a lot but to no use. Thanks in advance -Sandeep
