Hi,

I have problems making namenode and jobtracker remotely accessible.

It seems several people have had this problem before but I was 
unfortunately not able to find a solution yet.

I have a hadoop 0.20.6 cluster setup. All nodes with static IP
addresses, 
all wired up via short names, data0, data1, data2, master in /etc/hosts.

The master node hosts the name node as well as the job tracker. Both
listen
only to connection from the master node and will not accept remote 
connections:

> netstat -nltp

Proto Recv-Q Send-Q Local Address           Foreign Address
State       PID/Program name
tcp        0      0 127.0.0.1:3306          0.0.0.0:*
LISTEN      -               
tcp        0      0 0.0.0.0:10000           0.0.0.0:*
LISTEN      -               
tcp        0      0 0.0.0.0:22              0.0.0.0:*
LISTEN      -               
tcp6       0      0 a.b.c.d:60000       :::*                    LISTEN
19800/java      
tcp6       0      0 :::52038                :::*
LISTEN      19235/java      
tcp6       0      0 a.b.c.d:9000        :::*                    LISTEN
19235/java      
tcp6       0      0 a.b.c.d:9001        :::*                    LISTEN
19507/java      
tcp6       0      0 :::60010                :::*
LISTEN      19800/java      
tcp6       0      0 :::50090                :::*
LISTEN      19409/java      
tcp6       0      0 :::56429                :::*
LISTEN      19507/java      
tcp6       0      0 :::2222                 :::*
LISTEN      19717/java      
tcp6       0      0 :::50030                :::*
LISTEN      19507/java      
tcp6       0      0 :::38126                :::*
LISTEN      19409/java      
tcp6       0      0 :::80                   :::*
LISTEN      -               
tcp6       0      0 :::21                   :::*
LISTEN      -               
tcp6       0      0 :::50070                :::*
LISTEN      19235/java      
tcp6       0      0 :::22                   :::*
LISTEN      -               

(changed the real IP address to a.b.c.d). 

My hadoop/conf/core-site.xml looks like this:

<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!-- Put site-specific property overrides in this file. -->
<configuration>
<property>
        <name>fs.default.name</name>
        <value>hdfs://master:9000</value>
</property>
<property>
        <name>hadoop.tmp.dir</name>
        <value>/home/hadoop/data</value>
</property>
</configuration>

and hadoop/conf/mapred-site.xml like this:

<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!-- Put site-specific property overrides in this file. -->
<configuration>
<property>
  <name>mapred.job.tracker</name>
  <value>master:9001</value>
</property>
</configuration>


Using IP adresses rather than host names in core-site.xml or
hdfs-site.xml didn't
change anything (contrary to what other mailing list submissions
suggest).

Otherwise, the cluster starts up fine, all processes running, web
interfaces are reachable 
and report nothing unusual.

Any idea? I am blocked :-(

Thanks,
  Henning

Reply via email to