Hi,
I've just repeated the experiment. this is what I get:

NAMENODE

2009-09-14 17:08:25,304 INFO
org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting NameNode
STARTUP_MSG:   host = XXX/192.*.*.*
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 0.20.2-dev
STARTUP_MSG:   build =  -r ; compiled by 'vincenzo' on Mon Sep 14 15:49:43
CEST 2009
************************************************************/
2009-09-14 17:08:25,442 INFO org.apache.hadoop.ipc.metrics.RpcMetrics:
Initializing RPC Metrics with hostName=NameNode, port=8020
2009-09-14 17:08:25,448 INFO
org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up at:
XXX/192.*.*.*:8020
2009-09-14 17:08:25,450 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
Initializing JVM Metrics with processName=NameNode, sessionId=null
2009-09-14 17:08:25,453 INFO
org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics: Initializing
NameNodeMeterics using context
object:org.apache.hadoop.metrics.spi.NullContext
2009-09-14 17:08:25,530 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
fsOwner=vincenzo,vincenzo
2009-09-14 17:08:25,530 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup=supergroup
2009-09-14 17:08:25,530 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
isPermissionEnabled=true
2009-09-14 17:08:25,540 INFO
org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics:
Initializing FSNamesystemMetrics using context
object:org.apache.hadoop.metrics.spi.NullContext
2009-09-14 17:08:25,542 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered
FSNamesystemStatusMBean
2009-09-14 17:08:25,581 INFO org.apache.hadoop.hdfs.server.common.Storage:
Number of files = 1
2009-09-14 17:08:25,586 INFO org.apache.hadoop.hdfs.server.common.Storage:
Number of files under construction = 0
2009-09-14 17:08:25,586 INFO org.apache.hadoop.hdfs.server.common.Storage:
Image file of size 98 loaded in 0 seconds.
2009-09-14 17:08:25,586 INFO org.apache.hadoop.hdfs.server.common.Storage:
Edits file /tmp/hadoop-vincenzo/dfs/name/current/edits of size 4 edits # 0
loaded in 0 seconds.
2009-09-14 17:08:25,590 INFO org.apache.hadoop.hdfs.server.common.Storage:
Image file of size 98 saved in 0 seconds.
2009-09-14 17:08:25,602 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished loading
FSImage in 117 msecs
2009-09-14 17:08:25,603 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total number of blocks
= 0
2009-09-14 17:08:25,603 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of invalid
blocks = 0
2009-09-14 17:08:25,603 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
under-replicated blocks = 0
2009-09-14 17:08:25,603 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
over-replicated blocks = 0
2009-09-14 17:08:25,603 INFO org.apache.hadoop.hdfs.StateChange: STATE*
Leaving safe mode after 0 secs.
2009-09-14 17:08:25,604 INFO org.apache.hadoop.hdfs.StateChange: STATE*
Network topology has 0 racks and 0 datanodes
2009-09-14 17:08:25,604 INFO org.apache.hadoop.hdfs.StateChange: STATE*
UnderReplicatedBlocks has 0 blocks
2009-09-14 17:08:25,802 INFO org.mortbay.log: Logging to
org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
org.mortbay.log.Slf4jLog
2009-09-14 17:08:25,893 INFO org.apache.hadoop.http.HttpServer: Port
returned by webServer.getConnectors()[0].getLocalPort() before open() is -1.
Opening the listener on 50070
2009-09-14 17:08:25,894 INFO org.apache.hadoop.http.HttpServer:
listener.getLocalPort() returned 50070
webServer.getConnectors()[0].getLocalPort() returned 50070
2009-09-14 17:08:25,894 INFO org.apache.hadoop.http.HttpServer: Jetty bound
to port 50070
2009-09-14 17:08:25,894 INFO org.mortbay.log: jetty-6.1.14

DATANODE

2009-09-14 17:08:26,768 INFO
org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting DataNode
STARTUP_MSG:   host = XXX/192.*.*.*
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 0.20.2-dev
STARTUP_MSG:   build =  -r ; compiled by 'vincenzo' on Mon Sep 14 15:49:43
CEST 2009
************************************************************/

JOBTRACKER

2009-09-14 17:08:28,721 INFO org.apache.hadoop.mapred.JobTracker:
STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting JobTracker
STARTUP_MSG:   host = XXX/192.*.*.*
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 0.20.2-dev
STARTUP_MSG:   build =  -r ; compiled by 'vincenzo' on Mon Sep 14 15:49:43
CEST 2009
************************************************************/
2009-09-14 17:08:28,829 INFO org.apache.hadoop.mapred.JobTracker: Scheduler
configured with (memSizeForMapSlotOnJT, memSizeForReduceSlotOnJT,
limitMaxMemForMapTasks, limitMaxMemForReduceTasks) (-1, -1, -1, -1)
2009-09-14 17:08:28,875 INFO org.apache.hadoop.ipc.metrics.RpcMetrics:
Initializing RPC Metrics with hostName=JobTracker, port=8021
2009-09-14 17:08:28,948 INFO org.mortbay.log: Logging to
org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
org.mortbay.log.Slf4jLog
2009-09-14 17:08:29,114 INFO org.apache.hadoop.http.HttpServer: Port
returned by webServer.getConnectors()[0].getLocalPort() before open() is -1.
Opening the listener on 50030
2009-09-14 17:08:29,116 INFO org.apache.hadoop.http.HttpServer:
listener.getLocalPort() returned 50030
webServer.getConnectors()[0].getLocalPort() returned 50030
2009-09-14 17:08:29,116 INFO org.apache.hadoop.http.HttpServer: Jetty bound
to port 50030
2009-09-14 17:08:29,116 INFO org.mortbay.log: jetty-6.1.14

TASKTRACER

2009-09-14 17:08:30,028 INFO org.apache.hadoop.mapred.TaskTracker:
STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting TaskTracker
STARTUP_MSG:   host = XXX/192.*.*.*
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 0.20.2-dev
STARTUP_MSG:   build =  -r ; compiled by 'vincenzo' on Mon Sep 14 15:49:43
CEST 2009
************************************************************/
2009-09-14 17:08:30,240 INFO org.mortbay.log: Logging to
org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
org.mortbay.log.Slf4jLog
2009-09-14 17:08:30,399 INFO org.apache.hadoop.http.HttpServer: Port
returned by webServer.getConnectors()[0].getLocalPort() before open() is -1.
Opening the listener on 50060
2009-09-14 17:08:30,407 INFO org.apache.hadoop.http.HttpServer:
listener.getLocalPort() returned 50060
webServer.getConnectors()[0].getLocalPort() returned 50060
2009-09-14 17:08:30,407 INFO org.apache.hadoop.http.HttpServer: Jetty bound
to port 50060
2009-09-14 17:08:30,407 INFO org.mortbay.log: jetty-6.1.14

SECONDARY NAME NODE

2009-09-14 17:08:27,666 INFO
org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting SecondaryNameNode
STARTUP_MSG:   host = XXX/192.*.*.*
STARTUP_MSG:   args = []
STARTUP_MSG:   version = 0.20.2-dev
STARTUP_MSG:   build =  -r ; compiled by 'vincenzo' on Mon Sep 14 15:49:43
CEST 2009
************************************************************/
2009-09-14 17:08:27,738 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
Initializing JVM Metrics with processName=SecondaryNameNode, sessionId=null

______________________

I've noticed now that the namenode says that 0 namenode are availabe, but
the configuration is the one suggested by the hadoop tutorial and all the
conf files (included masters and slaves) of my single-node cluster setup are
specified with full name.
Do you have any idea about why this happens?
Again, thanks for your help


2009/9/14 Todd Lipcon <[email protected]>

> That's not an error - that just means that the daemon thread is waiting for
> a connection (IO event)
>
> The logs in $HADOOP_HOME/log/ are entirely empty? Both the .log and .out
> files? I find that hard to believe :)
>
> -Todd
>
> On Mon, Sep 14, 2009 at 7:57 AM, Vincenzo Gulisano <
> [email protected]> wrote:
>
> > Hi Todd,
> > thanks for your answer. I've already tried this solution. No error is
> > reported.
> > As the program remains in a "wait state", no error is detected.
> > I've seen that the error
> > "sun.nio.ch.EPollArrayWrapper. epollWait (native method)"
> > affects other old bugs of hadoop, but I couldn't solve mine.
> > Thanks again
> >
> >
> >
> >
> > 2009/9/14 Todd Lipcon <[email protected]>
> >
> > > Hi Vincenzo,
> > >
> > > Look at the log output of your daemons. My guess is that you'll find
> > > something pretty clear there.
> > >
> > > -Todd
> > >
> > > On Mon, Sep 14, 2009 at 7:46 AM, Vincenzo Gulisano <
> > > [email protected]> wrote:
> > >
> > > > Hi,
> > > > after a lot of unsuccessful attempts of running hadoop distributed
> file
> > > > system on my machine, I've located one possible error.
> > > > Maybe you have some ideas about what's going on.
> > > >
> > > > Experiment:
> > > > What I'm doing is simply executing start-all.sh and hadoop dfsadmin
> > > -report
> > > >
> > > > After the setup I can check that everything is working using:
> > > >
> > > > jps
> > > > ...
> > > > 17421 NameNode
> > > > 17519 DataNode
> > > > 17611 SecondaryNameNode
> > > > 17685 JobTracker
> > > > 17778 TaskTracker
> > > > 18425 Jps
> > > > ...
> > > >
> > > >
> > > >
> > > >
> > > > AND
> > > >
> > > > sudo netstat -plten | grep java
> > > > ...
> > > > tcp        0      0 127.0.0.1:54310         0.0.0.0:* LISTEN
>  1062
> > > >      346907      17421/java      (namenode)
> > > > tcp        0      0 127.0.0.1:54311         0.0.0.0:* LISTEN
>  1062
> > > >      347480      17685/java      (job tracker)
> > > >
> > > >
> > > >
> > > >
> > > > 2 things happen launching the application:
> > > > 1) The program waits and nothing happens (99% of the times)
> > > > 2) The program works but the report shows that the HDFS has some
> > problems
> > > >
> > > > Taking a look to the debug:
> > > >
> > > >
> > > > main:
> > > >
> > > >  [1] java.lang.Object.wait (native method)
> > > >  [2] java.lang.Object.wait (Object.java:485)
> > > >  [3] org.apache.hadoop.ipc.Client.call (Client.java:725)
> > > >  [4] org.apache.hadoop.ipc.RPC$Invoker.invoke (RPC.java:220)
> > > >  [5] $Proxy0.getProtocolVersion (null)
> > > >  [6] org.apache.hadoop.ipc.RPC.getProxy (RPC.java:359)
> > > >  [7] org.apache.hadoop.hdfs.DFSClient.createRPCNamenode
> > > > (DFSClient.java:105)
> > > >  [8] org.apache.hadoop.hdfs.DFSClient.<init> (DFSClient.java:208)
> > > >  [9] org.apache.hadoop.hdfs.DFSClient.<init> (DFSClient.java:169)
> > > >  [10] org.apache.hadoop.hdfs.DistributedFileSystem.initialize
> > > > (DistributedFileSystem.java:82)
> > > >  [11] org.apache.hadoop.fs.FileSystem.createFileSystem
> > > > (FileSystem.java:1,384)
> > > >  [12] org.apache.hadoop.fs.FileSystem.access$200 (FileSystem.java:66)
> > > >  [13] org.apache.hadoop.fs.FileSystem$Cache.get
> (FileSystem.java:1,399)
> > > >  [14] org.apache.hadoop.fs.FileSystem.get (FileSystem.java:199)
> > > >  [15] org.apache.hadoop.fs.FileSystem.get (FileSystem.java:96)
> > > >  [16] org.apache.hadoop.fs.FsShell.init (FsShell.java:85)
> > > >  [17] org.apache.hadoop.hdfs.tools.DFSAdmin.run (DFSAdmin.java:777)
> > > >  [18] org.apache.hadoop.util.ToolRunner.run (ToolRunner.java:65)
> > > >  [19] org.apache.hadoop.util.ToolRunner.run (ToolRunner.java:79)
> > > >  [20] org.apache.hadoop.hdfs.tools.DFSAdmin.main (DFSAdmin.java:858)
> > > >
> > > > IPC Client (47) connection to localhost/127.0.0.1:8020 from
> vincenzo:
> > > >  [1] sun.nio.ch.EPollArrayWrapper.epollWait (native method)
> > > >  [2] sun.nio.ch.EPollArrayWrapper.poll (EPollArrayWrapper.java:215)
> > > >  [3] sun.nio.ch.EPollSelectorImpl.doSelect
> (EPollSelectorImpl.java:65)
> > > >  [4] sun.nio.ch.SelectorImpl.lockAndDoSelect (SelectorImpl.java:69)
> > > >  [5] sun.nio.ch.SelectorImpl.select (SelectorImpl.java:80)
> > > >  [6] org.apache.hadoop.net.SocketIOWithTimeout$SelectorPool.select
> > > > (SocketIOWithTimeout.java:332)
> > > >  [7] org.apache.hadoop.net.SocketIOWithTimeout.doIO
> > > > (SocketIOWithTimeout.java:157)
> > > >  [8] org.apache.hadoop.net.SocketInputStream.read
> > > > (SocketInputStream.java:155)
> > > >  [9] org.apache.hadoop.net.SocketInputStream.read
> > > > (SocketInputStream.java:128)
> > > >  [10] java.io.FilterInputStream.read (FilterInputStream.java:116)
> > > >  [11] org.apache.hadoop.ipc.Client$Connection$PingInputStream.read
> > > > (Client.java:276)
> > > >  [12] java.io.BufferedInputStream.fill (BufferedInputStream.java:218)
> > > >  [13] java.io.BufferedInputStream.read (BufferedInputStream.java:237)
> > > >  [14] java.io.DataInputStream.readInt (DataInputStream.java:370)
> > > >  [15] org.apache.hadoop.ipc.Client$Connection.receiveResponse
> > > > (Client.java:501)
> > > >  [16] org.apache.hadoop.ipc.Client$Connection.run (Client.java:446)
> > > >
> > > > Have you any idea about why this can happen?
> > > >
> > > > I've tries also to telnet the host:port and it works. I've tried all
> > > > possible addresses in the configuration (localhost / 127.0.0.1 / name
> /
> > > > name.domain ).
> > > >
> > > > Any help is appreciated,
> > > > Thanks in advance
> > > >
> > > > Vincenzo
> > > >
> > >
> >
> >
> >
> > --
> > Vincenzo Massimiliano Gulisano
> > PhD student - UPM - Distributed System Lab.
> >
>



-- 
Vincenzo Massimiliano Gulisano
PhD student - UPM - Distributed System Lab.

Reply via email to