Hi,

First things to check...

1) Can you ping the machines from an external client machine.
2) /etc/hosts? Not a centralized DNS server? Is your client also in your 
/etc/hosts?
3) Do you only have one active NIC card?

And of course I'm assuming that when you say you have the cloud up, you can 
launch jobs on the namenode and they run on all of the nodes?

-Mike

> Subject: namenode and jobtracker remote access problem
> From: [email protected]
> To: [email protected]
> Date: Fri, 5 Nov 2010 12:23:30 +0100
> 
> Hi,
> 
> I have problems making namenode and jobtracker remotely accessible.
> 
> It seems several people have had this problem before but I was 
> unfortunately not able to find a solution yet.
> 
> I have a hadoop 0.20.6 cluster setup. All nodes with static IP
> addresses, 
> all wired up via short names, data0, data1, data2, master in /etc/hosts.
> 
> The master node hosts the name node as well as the job tracker. Both
> listen
> only to connection from the master node and will not accept remote 
> connections:
> 
> > netstat -nltp
> 
> Proto Recv-Q Send-Q Local Address           Foreign Address
> State       PID/Program name
> tcp        0      0 127.0.0.1:3306          0.0.0.0:*
> LISTEN      -               
> tcp        0      0 0.0.0.0:10000           0.0.0.0:*
> LISTEN      -               
> tcp        0      0 0.0.0.0:22              0.0.0.0:*
> LISTEN      -               
> tcp6       0      0 a.b.c.d:60000       :::*                    LISTEN
> 19800/java      
> tcp6       0      0 :::52038                :::*
> LISTEN      19235/java      
> tcp6       0      0 a.b.c.d:9000        :::*                    LISTEN
> 19235/java      
> tcp6       0      0 a.b.c.d:9001        :::*                    LISTEN
> 19507/java      
> tcp6       0      0 :::60010                :::*
> LISTEN      19800/java      
> tcp6       0      0 :::50090                :::*
> LISTEN      19409/java      
> tcp6       0      0 :::56429                :::*
> LISTEN      19507/java      
> tcp6       0      0 :::2222                 :::*
> LISTEN      19717/java      
> tcp6       0      0 :::50030                :::*
> LISTEN      19507/java      
> tcp6       0      0 :::38126                :::*
> LISTEN      19409/java      
> tcp6       0      0 :::80                   :::*
> LISTEN      -               
> tcp6       0      0 :::21                   :::*
> LISTEN      -               
> tcp6       0      0 :::50070                :::*
> LISTEN      19235/java      
> tcp6       0      0 :::22                   :::*
> LISTEN      -               
> 
> (changed the real IP address to a.b.c.d). 
> 
> My hadoop/conf/core-site.xml looks like this:
> 
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> <!-- Put site-specific property overrides in this file. -->
> <configuration>
> <property>
>         <name>fs.default.name</name>
>         <value>hdfs://master:9000</value>
> </property>
> <property>
>         <name>hadoop.tmp.dir</name>
>         <value>/home/hadoop/data</value>
> </property>
> </configuration>
> 
> and hadoop/conf/mapred-site.xml like this:
> 
> <?xml version="1.0"?>
> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> <!-- Put site-specific property overrides in this file. -->
> <configuration>
> <property>
>   <name>mapred.job.tracker</name>
>   <value>master:9001</value>
> </property>
> </configuration>
> 
> 
> Using IP adresses rather than host names in core-site.xml or
> hdfs-site.xml didn't
> change anything (contrary to what other mailing list submissions
> suggest).
> 
> Otherwise, the cluster starts up fine, all processes running, web
> interfaces are reachable 
> and report nothing unusual.
> 
> Any idea? I am blocked :-(
> 
> Thanks,
>   Henning
> 
                                          

Reply via email to