Well...

0.0.0.0 means that its listening on all networks and in your case... eth0 and 
127.0.0.1.

I'd try adding your client to the /etc/hosts on the machines.



> Subject: RE: namenode and jobtracker remote access problem
> From: [email protected]
> To: [email protected]
> Date: Fri, 5 Nov 2010 13:04:19 +0100
> 
> Hi Mike,
> 
> 1) yes. My client can ssh into any of nodes.
> 2) No, unfortunately not (hosted machines, no domain yet, just IP
> addresses). My client is not in /etc/hosts of any of the nodes. Why?
> Would they do reverse lookups?
> 3) looking at ifconfig's output there is only eth0 and lo. So I assume
> that is a yes to your question.
> 
> My wild guess is that the namenode (and jobtracker) code by default try
> to resolve the host name specified in fs.default.name and
> mapred.job.tracker resp. 
> and use the resulting IP to open the server socket (or channel). Rather
> than 0.0.0.0. 
> 
> But if that was the case, many, really many people should have the same
> problem....
> 
> Thanks,
>   Henning
> 
> 
> Am Freitag, den 05.11.2010, 06:55 -0500 schrieb Michael Segel:
> 
> > Hi,
> > 
> > First things to check...
> > 
> > 1) Can you ping the machines from an external client machine.
> > 2) /etc/hosts? Not a centralized DNS server? Is your client also in your 
> > /etc/hosts?
> > 3) Do you only have one active NIC card?
> > 
> > And of course I'm assuming that when you say you have the cloud up, you can 
> > launch jobs on the namenode and they run on all of the nodes?
> > 
> > -Mike
> > 
> > > Subject: namenode and jobtracker remote access problem
> > > From: [email protected]
> > > To: [email protected]
> > > Date: Fri, 5 Nov 2010 12:23:30 +0100
> > > 
> > > Hi,
> > > 
> > > I have problems making namenode and jobtracker remotely accessible.
> > > 
> > > It seems several people have had this problem before but I was 
> > > unfortunately not able to find a solution yet.
> > > 
> > > I have a hadoop 0.20.6 cluster setup. All nodes with static IP
> > > addresses, 
> > > all wired up via short names, data0, data1, data2, master in /etc/hosts.
> > > 
> > > The master node hosts the name node as well as the job tracker. Both
> > > listen
> > > only to connection from the master node and will not accept remote 
> > > connections:
> > > 
> > > > netstat -nltp
> > > 
> > > Proto Recv-Q Send-Q Local Address           Foreign Address
> > > State       PID/Program name
> > > tcp        0      0 127.0.0.1:3306          0.0.0.0:*
> > > LISTEN      -               
> > > tcp        0      0 0.0.0.0:10000           0.0.0.0:*
> > > LISTEN      -               
> > > tcp        0      0 0.0.0.0:22              0.0.0.0:*
> > > LISTEN      -               
> > > tcp6       0      0 a.b.c.d:60000       :::*                    LISTEN
> > > 19800/java      
> > > tcp6       0      0 :::52038                :::*
> > > LISTEN      19235/java      
> > > tcp6       0      0 a.b.c.d:9000        :::*                    LISTEN
> > > 19235/java      
> > > tcp6       0      0 a.b.c.d:9001        :::*                    LISTEN
> > > 19507/java      
> > > tcp6       0      0 :::60010                :::*
> > > LISTEN      19800/java      
> > > tcp6       0      0 :::50090                :::*
> > > LISTEN      19409/java      
> > > tcp6       0      0 :::56429                :::*
> > > LISTEN      19507/java      
> > > tcp6       0      0 :::2222                 :::*
> > > LISTEN      19717/java      
> > > tcp6       0      0 :::50030                :::*
> > > LISTEN      19507/java      
> > > tcp6       0      0 :::38126                :::*
> > > LISTEN      19409/java      
> > > tcp6       0      0 :::80                   :::*
> > > LISTEN      -               
> > > tcp6       0      0 :::21                   :::*
> > > LISTEN      -               
> > > tcp6       0      0 :::50070                :::*
> > > LISTEN      19235/java      
> > > tcp6       0      0 :::22                   :::*
> > > LISTEN      -               
> > > 
> > > (changed the real IP address to a.b.c.d). 
> > > 
> > > My hadoop/conf/core-site.xml looks like this:
> > > 
> > > <?xml version="1.0"?>
> > > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> > > <!-- Put site-specific property overrides in this file. -->
> > > <configuration>
> > > <property>
> > >         <name>fs.default.name</name>
> > >         <value>hdfs://master:9000</value>
> > > </property>
> > > <property>
> > >         <name>hadoop.tmp.dir</name>
> > >         <value>/home/hadoop/data</value>
> > > </property>
> > > </configuration>
> > > 
> > > and hadoop/conf/mapred-site.xml like this:
> > > 
> > > <?xml version="1.0"?>
> > > <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
> > > <!-- Put site-specific property overrides in this file. -->
> > > <configuration>
> > > <property>
> > >   <name>mapred.job.tracker</name>
> > >   <value>master:9001</value>
> > > </property>
> > > </configuration>
> > > 
> > > 
> > > Using IP adresses rather than host names in core-site.xml or
> > > hdfs-site.xml didn't
> > > change anything (contrary to what other mailing list submissions
> > > suggest).
> > > 
> > > Otherwise, the cluster starts up fine, all processes running, web
> > > interfaces are reachable 
> > > and report nothing unusual.
> > > 
> > > Any idea? I am blocked :-(
> > > 
> > > Thanks,
> > >   Henning
> > > 
> >                                       
                                          

Reply via email to