Make sure your hosts file on your namenode is setup correctly:

127.0.0.1               localhost.localdomain localhost
10.x.x.x                myhostname.mydomain.com myhostname

As opposed to:

127.0.0.1 localhost.localdomain localhost myhostname.mydomain.com myhostname

The problem may be that the machine is listening on only the local interface. If you do a ping myhostname from the local box you should receive the real IP and not the loopback address.

Let me know if this was the problem or if you need more help.

Dennis Kubes

cybercouf wrote:
I'm trying to setup hadoop using these guides:
http://wiki.apache.org/nutch/NutchHadoopTutorial and
http://www.nabble.com/Nutch-Step-by-Step-Maybe-someone-will-find-this-useful---tf3526281.html

But i'm stuck at the early step: having a single machine running.
using nutch 0.8.1 and so the provided hadoop "hadoop-0.4.0-patched.jar"
JVM sun 1.5.0_11

When I start the namenode (using ./bin/start-all.sh) I have this in the
namenode-log:

2007-05-02 12:39:51,335 INFO  util.Credential - Checking Resource aliases
2007-05-02 12:39:51,349 INFO  http.HttpServer - Version Jetty/5.1.4
2007-05-02 12:39:51,350 WARN  servlet.WebApplicationContext - Web
application not found
/home/nutch/search/file:/home/nutch/search/lib/hadoop-0.4.0-patched.jar!/webapps/dfs
2007-05-02 12:39:51,351 WARN  servlet.WebApplicationContext - Configuration
error on
/home/nutch/search/file:/home/nutch/search/lib/hadoop-0.4.0-patched.jar!/webapps/dfs
java.io.FileNotFoundException:
/home/nutch/search/file:/home/nutch/search/lib/hadoop-0.4.0-patched.jar!/webapps/dfs
        at
org.mortbay.jetty.servlet.WebApplicationContext.resolveWebApp(WebApplicationContext.java:266)
        at
org.mortbay.jetty.servlet.WebApplicationContext.doStart(WebApplicationContext.java:449)
        at org.mortbay.util.Container.start(Container.java:72)
        at org.mortbay.http.HttpServer.doStart(HttpServer.java:753)
        at org.mortbay.util.Container.start(Container.java:72)
        at
org.apache.hadoop.mapred.StatusHttpServer.start(StatusHttpServer.java:138)
        at org.apache.hadoop.dfs.FSNamesystem.<init>(FSNamesystem.java:173)
        at org.apache.hadoop.dfs.NameNode.<init>(NameNode.java:91)
        at org.apache.hadoop.dfs.NameNode.<init>(NameNode.java:82)
        at org.apache.hadoop.dfs.NameNode.main(NameNode.java:491)
2007-05-02 12:39:51,353 INFO  util.Container - Started
HttpContext[/logs,/logs]
2007-05-02 12:39:51,353 INFO  util.Container - Started
HttpContext[/static,/static]
2007-05-02 12:39:51,357 INFO  http.SocketListener - Started SocketListener
on 0.0.0.0:50070

and after I can't access it:
$ ./bin/hadoop dfs -ls
ls: Connection refused

hadoop.log:
2007-05-02 12:41:40,030 WARN  fs.DFSClient - Problem renewing lease for
DFSClient_2015604182: java.net.ConnectException: Connection refused
at java.net.PlainSocketImpl.socketConnect(Native Method) [...]



1. I can't understant why there is this FileNotFound execption, I didn't
change anything in the hadoop nutch jar file.

2. It looks like the namenode is running (when I stop it I have the message
"stopping namenode"), but why I can't access it ? (is this ip from the log
correct? 0.0.0.0:50070)
all is on the same machine, and my conf file looks ok:
fs.default.name   myhostname:9000
mapred.job.tracker  myhostname:9001
mapred.map.tasks  2
mapred.reduce.tasks  2
dfs.name.dir  /home/nutch/filesystem/name
dfs.data.dir  /home/nutch/filesystem/data
mapred.system.dir  /home/nutch/filesystem/mapreduce/system
mapred.local.dir  /home/nutch/filesystem/mapreduce/local
dfs.replication  1

Reply via email to