Make sure your hosts file on your namenode is setup correctly:
127.0.0.1 localhost.localdomain localhost
10.x.x.x myhostname.mydomain.com myhostname
As opposed to:
127.0.0.1 localhost.localdomain localhost
myhostname.mydomain.com myhostname
The problem may be that the machine is listening on only the local
interface. If you do a ping myhostname from the local box you should
receive the real IP and not the loopback address.
Let me know if this was the problem or if you need more help.
Dennis Kubes
cybercouf wrote:
I'm trying to setup hadoop using these guides:
http://wiki.apache.org/nutch/NutchHadoopTutorial and
http://www.nabble.com/Nutch-Step-by-Step-Maybe-someone-will-find-this-useful---tf3526281.html
But i'm stuck at the early step: having a single machine running.
using nutch 0.8.1 and so the provided hadoop "hadoop-0.4.0-patched.jar"
JVM sun 1.5.0_11
When I start the namenode (using ./bin/start-all.sh) I have this in the
namenode-log:
2007-05-02 12:39:51,335 INFO util.Credential - Checking Resource aliases
2007-05-02 12:39:51,349 INFO http.HttpServer - Version Jetty/5.1.4
2007-05-02 12:39:51,350 WARN servlet.WebApplicationContext - Web
application not found
/home/nutch/search/file:/home/nutch/search/lib/hadoop-0.4.0-patched.jar!/webapps/dfs
2007-05-02 12:39:51,351 WARN servlet.WebApplicationContext - Configuration
error on
/home/nutch/search/file:/home/nutch/search/lib/hadoop-0.4.0-patched.jar!/webapps/dfs
java.io.FileNotFoundException:
/home/nutch/search/file:/home/nutch/search/lib/hadoop-0.4.0-patched.jar!/webapps/dfs
at
org.mortbay.jetty.servlet.WebApplicationContext.resolveWebApp(WebApplicationContext.java:266)
at
org.mortbay.jetty.servlet.WebApplicationContext.doStart(WebApplicationContext.java:449)
at org.mortbay.util.Container.start(Container.java:72)
at org.mortbay.http.HttpServer.doStart(HttpServer.java:753)
at org.mortbay.util.Container.start(Container.java:72)
at
org.apache.hadoop.mapred.StatusHttpServer.start(StatusHttpServer.java:138)
at org.apache.hadoop.dfs.FSNamesystem.<init>(FSNamesystem.java:173)
at org.apache.hadoop.dfs.NameNode.<init>(NameNode.java:91)
at org.apache.hadoop.dfs.NameNode.<init>(NameNode.java:82)
at org.apache.hadoop.dfs.NameNode.main(NameNode.java:491)
2007-05-02 12:39:51,353 INFO util.Container - Started
HttpContext[/logs,/logs]
2007-05-02 12:39:51,353 INFO util.Container - Started
HttpContext[/static,/static]
2007-05-02 12:39:51,357 INFO http.SocketListener - Started SocketListener
on 0.0.0.0:50070
and after I can't access it:
$ ./bin/hadoop dfs -ls
ls: Connection refused
hadoop.log:
2007-05-02 12:41:40,030 WARN fs.DFSClient - Problem renewing lease for
DFSClient_2015604182: java.net.ConnectException: Connection refused
at java.net.PlainSocketImpl.socketConnect(Native Method)
[...]
1. I can't understant why there is this FileNotFound execption, I didn't
change anything in the hadoop nutch jar file.
2. It looks like the namenode is running (when I stop it I have the message
"stopping namenode"), but why I can't access it ? (is this ip from the log
correct? 0.0.0.0:50070)
all is on the same machine, and my conf file looks ok:
fs.default.name myhostname:9000
mapred.job.tracker myhostname:9001
mapred.map.tasks 2
mapred.reduce.tasks 2
dfs.name.dir /home/nutch/filesystem/name
dfs.data.dir /home/nutch/filesystem/data
mapred.system.dir /home/nutch/filesystem/mapreduce/system
mapred.local.dir /home/nutch/filesystem/mapreduce/local
dfs.replication 1