Hi Mike,

1) yes. My client can ssh into any of nodes.
2) No, unfortunately not (hosted machines, no domain yet, just IP
addresses). My client is not in /etc/hosts of any of the nodes. Why?
Would they do reverse lookups?
3) looking at ifconfig's output there is only eth0 and lo. So I assume
that is a yes to your question.

My wild guess is that the namenode (and jobtracker) code by default try
to resolve the host name specified in fs.default.name and
mapred.job.tracker resp. 
and use the resulting IP to open the server socket (or channel). Rather
than 0.0.0.0. 

But if that was the case, many, really many people should have the same
problem....

Thanks,
  Henning


Am Freitag, den 05.11.2010, 06:55 -0500 schrieb Michael Segel:

> Hi,
> 
> First things to check...
> 
> 1) Can you ping the machines from an external client machine.
> 2) /etc/hosts? Not a centralized DNS server? Is your client also in your 
> /etc/hosts?
> 3) Do you only have one active NIC card?
> 
> And of course I'm assuming that when you say you have the cloud up, you can 
> launch jobs on the namenode and they run on all of the nodes?
> 
> -Mike
> 
> > Subject: namenode and jobtracker remote access problem
> > From: [email protected]
> > To: [email protected]
> > Date: Fri, 5 Nov 2010 12:23:30 +0100
> > 
> > Hi,
> > 
> > I have problems making namenode and jobtracker remotely accessible.
> > 
> > It seems several people have had this problem before but I was 
> > unfortunately not able to find a solution yet.
> > 
> > I have a hadoop 0.20.6 cluster setup. All nodes with static IP
> > addresses, 
> > all wired up via short names, data0, data1, data2, master in /etc/hosts.
> > 
> > The master node hosts the name node as well as the job tracker. Both
> > listen
> > only to connection from the master node and will not accept remote 
> > connections:
> > 
> > > netstat -nltp
> > 
> > Proto Recv-Q Send-Q Local Address           Foreign Address
> > State       PID/Program name
> > tcp        0      0 127.0.0.1:3306          0.0.0.0:*
> > LISTEN      -               
> > tcp        0      0 0.0.0.0:10000           0.0.0.0:*
> > LISTEN      -               
> > tcp        0      0 0.0.0.0:22              0.0.0.0:*
> > LISTEN      -               
> > tcp6       0      0 a.b.c.d:60000       :::*                    LISTEN
> > 19800/java      
> > tcp6       0      0 :::52038                :::*
> > LISTEN      19235/java      
> > tcp6       0      0 a.b.c.d:9000        :::*                    LISTEN
> > 19235/java      
> > tcp6       0      0 a.b.c.d:9001        :::*                    LISTEN
> > 19507/java      
> > tcp6       0      0 :::60010                :::*
> > LISTEN      19800/java      
> > tcp6       0      0 :::50090                :::*
> > LISTEN      19409/java      
> > tcp6       0      0 :::56429                :::*
> > LISTEN      19507/java      
> > tcp6       0      0 :::2222                 :::*
> > LISTEN      19717/java      
> > tcp6       0      0 :::50030                :::*
> > LISTEN      19507/java      
> > tcp6       0      0 :::38126                :::*
> > LISTEN      19409/java      
> > tcp6       0      0 :::80                   :::*
> > LISTEN      -               
> > tcp6       0      0 :::21                   :::*
> > LISTEN      -               
> > tcp6       0      0 :::50070                :::*
> > LISTEN      19235/java      
> > tcp6       0      0 :::22                   :::*
> > LISTEN      -               
> > 
> > (changed the real IP address to a.b.c.d). 
> > 
> > My hadoop/conf/core-site.xml looks like this:
> > 
> > <?xml version="1.0"?>
> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> > <!-- Put site-specific property overrides in this file. -->
> > <configuration>
> > <property>
> >         <name>fs.default.name</name>
> >         <value>hdfs://master:9000</value>
> > </property>
> > <property>
> >         <name>hadoop.tmp.dir</name>
> >         <value>/home/hadoop/data</value>
> > </property>
> > </configuration>
> > 
> > and hadoop/conf/mapred-site.xml like this:
> > 
> > <?xml version="1.0"?>
> > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> > <!-- Put site-specific property overrides in this file. -->
> > <configuration>
> > <property>
> >   <name>mapred.job.tracker</name>
> >   <value>master:9001</value>
> > </property>
> > </configuration>
> > 
> > 
> > Using IP adresses rather than host names in core-site.xml or
> > hdfs-site.xml didn't
> > change anything (contrary to what other mailing list submissions
> > suggest).
> > 
> > Otherwise, the cluster starts up fine, all processes running, web
> > interfaces are reachable 
> > and report nothing unusual.
> > 
> > Any idea? I am blocked :-(
> > 
> > Thanks,
> >   Henning
> > 
>                                         

Reply via email to