On 05/19/2011 10:35 PM, 李�S wrote:
> Hi Marcos,
> Thanks for your reply.
> The temporary directory '/tmp/hadoop-xxx' is defined in hadoop core
> jar's configuration file "*core-default.xml*". Do u think this may
> cause the failure? Bellow is the detail config:
>
> <property>
> <name>hadoop.tmp.dir</name>
> <value>/tmp/hadoop-${user.name}</value>
> <description>A base for other temporary directories.</description>
> </property>
>
> And what's the other config files do u need? Almostly, I didn't modify
> any configuration after downloading the hadoop-0.20.2 files, I think
> those configuration are all the default values.
Yes, those are the default values, but I think that you can test with
another directory because this is a temporary directory , and it can be
erased easy.
For example, when you use the CDH3, the default value there is
/var/lib/hadoop-0.20.2/cache/${user.name}, which is more convenient.
Of course, it's a recommendation.
You can search the Lars Francke's Blog (http://blog.lars-francke.de/)
where he did a excellent work explaining the manual installation of a
Hadoop Cluster.
Regards
> 2011-05-20
> ------------------------------------------------------------------------
> 李�S
> ------------------------------------------------------------------------
> *发件人:* Marcos Ortiz
> *发送时间:* 2011-05-19 20:40:06
> *收件人:* mapreduce-user
> *抄送:* 李�S
> *主题:* Re: run hadoop pseudo-distribute examples failed
> On 05/18/2011 10:53 PM, 李�S wrote:
>> Hi All,
>> I'm trying to run hadoop(0.20.2) examples in Pseudo-Distributed Mode
>> following the hadoop user guide. After I run the 'start-all.sh', it
>> seems the namenode can't connect to datanode.
>> 'SSH localhost' is OK on my server. Someone advises to rm
>> '/tmp/hadoop-XXXX' and format namenode again, but it doesn't work.
>> And 'iptables -L' shows there is no firewall rules in my server:
>>
>> test:/home/liyun2010# iptables -L
>> Chain INPUT (policy ACCEPT)
>> target prot opt source destination
>> Chain FORWARD (policy ACCEPT)
>> target prot opt source destination
>> Chain OUTPUT (policy ACCEPT)
>> target prot opt source destination
>>
>> Is there anyone can give me more advice? Thanks!
>> Bellow is my namenode and datanode log files:
>> liyun2010@test:~/hadoop-0.20.2/logs$
>> <mailto:liyun2010@test:%7E/hadoop-0.20.2/logs$> cat
>> hadoop-liyun2010-namenode-test.puppet.com.log
>>
>> 2011-05-19 10:58:25,938 INFO
>> org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG:
>> /************************************************************
>> STARTUP_MSG: Starting NameNode
>> STARTUP_MSG: host = test.puppet.com/127.0.0.1
>> STARTUP_MSG: args = []
>> STARTUP_MSG: version = 0.20.2
>> STARTUP_MSG: build =
>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20
>> -r 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010
>> ************************************************************/
>> 2011-05-19 10:58:26,197 INFO
>> org.apache.hadoop.ipc.metrics.RpcMetrics: Initializing RPC
>> Metrics with hostName=NameNode, port=9000
>> 2011-05-19 10:58:26,212 INFO
>> org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up at:
>> test.puppet.com/127.0.0.1:9000
>> 2011-05-19 10:58:26,220 INFO
>> org.apache.hadoop.metrics.jvm.JvmMetrics: Initializing JVM
>> Metrics with processName=NameNode, sessionId=null
>> 2011-05-19 10:58:26,224 INFO
>> org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics:
>> Initializing NameNodeMeterics using context
>> object:org.apache.hadoop.metrics.spi.NullContext
>> 2011-05-19 10:58:26,405 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>> fsOwner=liyun2010,users
>> 2011-05-19 10:58:26,406 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>> supergroup=supergroup
>> 2011-05-19 10:58:26,406 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>> isPermissionEnabled=true
>> 2011-05-19 10:58:26,429 INFO
>> org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics:
>> Initializing FSNamesystemMetrics using context
>> object:org.apache.hadoop.metrics.spi.NullContext
>> 2011-05-19 10:58:26,434 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered
>> FSNamesystemStatusMBean
>> 2011-05-19 10:58:26,511 INFO
>> org.apache.hadoop.hdfs.server.common.Storage: Number of files = 9
>> 2011-05-19 10:58:26,524 INFO
>> org.apache.hadoop.hdfs.server.common.Storage: Number of files
>> under construction = 1
>> 2011-05-19 10:58:26,530 INFO
>> org.apache.hadoop.hdfs.server.common.Storage: Image file of size
>> 920 loaded in 0 seconds.
>> 2011-05-19 10:58:26,606 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Invalid
>> opcode, reached end of edit log Number of transactions found 99
>> 2011-05-19 10:58:26,606 INFO
>> org.apache.hadoop.hdfs.server.common.Storage: Edits file
>> /tmp/hadoop-liyun2010/dfs/name/current/edits of size 1049092
>> edits # 99 loaded in 0 seconds.
>> 2011-05-19 10:58:26,660 INFO
>> org.apache.hadoop.hdfs.server.common.Storage: Image file of size
>> 920 saved in 0 seconds.
>> 2011-05-19 10:58:26,810 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished
>> loading FSImage in 505 msecs
>> 2011-05-19 10:58:26,825 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total number
>> of blocks = 0
>> 2011-05-19 10:58:26,825 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
>> invalid blocks = 0
>> 2011-05-19 10:58:26,825 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
>> under-replicated blocks = 0
>> 2011-05-19 10:58:26,825 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
>> over-replicated blocks = 0
>> 2011-05-19 10:58:26,825 INFO org.apache.hadoop.hdfs.StateChange:
>> STATE* Leaving safe mode after 0 secs.
>> 2011-05-19 10:58:26,826 INFO org.apache.hadoop.hdfs.StateChange:
>> STATE* Network topology has 0 racks and 0 datanodes
>> 2011-05-19 10:58:26,826 INFO org.apache.hadoop.hdfs.StateChange:
>> STATE* UnderReplicatedBlocks has 0 blocks
>> 2011-05-19 10:58:27,025 INFO org.mortbay.log: Logging to
>> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
>> org.mortbay.log.Slf4jLog
>> 2011-05-19 10:58:27,174 INFO org.apache.hadoop.http.HttpServer:
>> Port returned by webServer.getConnectors()[0].getLocalPort()
>> before open() is -1. Opening the listener on 50070
>> 2011-05-19 10:58:27,178 INFO org.apache.hadoop.http.HttpServer:
>> listener.getLocalPort() returned 50070
>> webServer.getConnectors()[0].getLocalPort() returned 50070
>> 2011-05-19 10:58:27,178 INFO org.apache.hadoop.http.HttpServer:
>> Jetty bound to port 50070
>> 2011-05-19 10:58:27,179 INFO org.mortbay.log: jetty-6.1.14
>> 2011-05-19 10:58:27,269 WARN org.mortbay.log: Can't reuse
>> /tmp/Jetty_0_0_0_0_50070_hdfs____w2cu08, using
>> /tmp/Jetty_0_0_0_0_50070_hdfs____w2cu08_740365192444258489
>> 2011-05-19 10:58:28,610 INFO org.mortbay.log: Started
>> [email protected]:50070
>> 2011-05-19 10:58:28,611 INFO
>> org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up
>> at: 0.0.0.0:50070
>> 2011-05-19 10:58:28,612 INFO org.apache.hadoop.ipc.Server: IPC
>> Server Responder: starting
>> 2011-05-19 10:58:28,613 INFO org.apache.hadoop.ipc.Server: IPC
>> Server listener on 9000: starting
>> 2011-05-19 10:58:28,617 INFO org.apache.hadoop.ipc.Server: IPC
>> Server handler 0 on 9000: starting
>> 2011-05-19 10:58:28,618 INFO org.apache.hadoop.ipc.Server: IPC
>> Server handler 1 on 9000: starting
>> 2011-05-19 10:58:28,621 INFO org.apache.hadoop.ipc.Server: IPC
>> Server handler 2 on 9000: starting
>> 2011-05-19 10:58:28,625 INFO org.apache.hadoop.ipc.Server: IPC
>> Server handler 4 on 9000: starting
>> 2011-05-19 10:58:28,625 INFO org.apache.hadoop.ipc.Server: IPC
>> Server handler 5 on 9000: starting
>> 2011-05-19 10:58:28,626 INFO org.apache.hadoop.ipc.Server: IPC
>> Server handler 6 on 9000: starting
>> 2011-05-19 10:58:28,627 INFO org.apache.hadoop.ipc.Server: IPC
>> Server handler 3 on 9000: starting
>> 2011-05-19 10:58:28,629 INFO org.apache.hadoop.ipc.Server: IPC
>> Server handler 8 on 9000: starting
>> 2011-05-19 10:58:28,630 INFO org.apache.hadoop.ipc.Server: IPC
>> Server handler 9 on 9000: starting
>> 2011-05-19 10:58:28,630 INFO org.apache.hadoop.ipc.Server: IPC
>> Server handler 7 on 9000: starting
>> 2011-05-19 10:58:30,680 INFO org.apache.hadoop.hdfs.StateChange:
>> BLOCK* NameSystem.registerDatanode: node registration from
>> 127.0.0.1:50010 storage DS-173493047-127.0.0.1-50010-1305278767521
>> 2011-05-19 10:58:30,687 INFO
>> org.apache.hadoop.net.NetworkTopology: Adding a new node:
>> /default-rack/127.0.0.1:50010
>> 2011-05-19 10:58:39,361 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=liyun2010,users ip=/127.0.0.1 cmd=listStatus
>> src=/tmp/hadoop-liyun2010/mapred/system dst=nullperm=null
>> 2011-05-19 10:58:39,393 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=liyun2010,users ip=/127.0.0.1 cmd=delete
>> src=/tmp/hadoop-liyun2010/mapred/system dst=nullperm=null
>> 2011-05-19 10:58:39,405 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=liyun2010,users ip=/127.0.0.1 cmd=mkdirs
>> src=/tmp/hadoop-liyun2010/mapred/system
>> dst=nullperm=liyun2010:supergroup:rwxr-xr-x
>> 2011-05-19 10:58:39,417 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=liyun2010,users ip=/127.0.0.1 cmd=setPermission
>> src=/tmp/hadoop-liyun2010/mapred/system
>> dst=nullperm=liyun2010:supergroup:rwx-wx-wx
>> 2011-05-19 10:58:39,507 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=liyun2010,users ip=/127.0.0.1 cmd=create
>> src=/tmp/hadoop-liyun2010/mapred/system/jobtracker.infodst=null
>> perm=liyun2010:supergroup:rw-r--r--
>> 2011-05-19 10:58:39,530 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=liyun2010,users ip=/127.0.0.1 cmd=setPermission
>> src=/tmp/hadoop-liyun2010/mapred/system/jobtracker.info dst=null
>> perm=liyun2010:supergroup:rw-------
>> 2011-05-19 10:58:39,538 WARN
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Not able to
>> place enough replicas, still in need of 1
>> 2011-05-19 10:58:39,541 INFO org.apache.hadoop.ipc.Server: IPC
>> Server handler 7 on 9000, call
>> addBlock(/tmp/hadoop-liyun2010/mapred/system/jobtracker.info,
>> DFSClient_1143649887) from 127.0.0.1:56940: error:
>> java.io.IOException: File
>> /tmp/hadoop-liyun2010/mapred/system/jobtracker.info could only be
>> replicated to 0 nodes, instead of 1
>> java.io.IOException: File
>> /tmp/hadoop-liyun2010/mapred/system/jobtracker.info could only be
>> replicated to 0 nodes, instead of 1
>> at
>>
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1271)
>> at
>>
>> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422)
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> at
>>
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> at
>>
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> at java.lang.reflect.Method.invoke(Method.java:597)
>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)
>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)
>> at java.security.AccessController.doPrivileged(Native Method)
>> at javax.security.auth.Subject.doAs(Subject.java:396)
>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)
>> 2011-05-19 10:58:39,554 INFO
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>> ugi=liyun2010,users ip=/127.0.0.1 cmd=delete
>> src=/tmp/hadoop-liyun2010/mapred/system/jobtracker.infodst=null
>> perm=null
>>
>> liyun2010@test:~/hadoop-0.20.2/logs$ cat
>> hadoop-liyun2010-datanode-test.puppet.com.log
>> 2011-05-19 10:58:27,372 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG:
>> /************************************************************
>> STARTUP_MSG: Starting DataNode
>> STARTUP_MSG: host = test.puppet.com/127.0.0.1
>> STARTUP_MSG: args = []
>> STARTUP_MSG: version = 0.20.2
>> STARTUP_MSG: build =
>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20
>> -r 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010
>> ************************************************************/
>> 2011-05-19 10:58:28,932 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Registered
>> FSDatasetStatusMBean
>> 2011-05-19 10:58:28,938 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Opened info server
>> at 50010
>> 2011-05-19 10:58:28,942 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Balancing bandwith
>> is 1048576 bytes/s
>> 2011-05-19 10:58:29,137 INFO org.mortbay.log: Logging to
>> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
>> org.mortbay.log.Slf4jLog
>> 2011-05-19 10:58:29,341 INFO org.apache.hadoop.http.HttpServer: Port
>> returned by webServer.getConnectors()[0].getLocalPort() before open()
>> is -1. Opening the listener on 50075
>> 2011-05-19 10:58:29,342 INFO org.apache.hadoop.http.HttpServer:
>> listener.getLocalPort() returned 50075
>> webServer.getConnectors()[0].getLocalPort() returned 50075
>> 2011-05-19 10:58:29,342 INFO org.apache.hadoop.http.HttpServer: Jetty
>> bound to port 50075
>> 2011-05-19 10:58:29,342 INFO org.mortbay.log: jetty-6.1.14
>> 2011-05-19 10:58:30,600 INFO org.mortbay.log: Started
>> [email protected]:50075
>> 2011-05-19 10:58:30,620 INFO
>> org.apache.hadoop.metrics.jvm.JvmMetrics: Initializing JVM Metrics
>> with processName=DataNode, sessionId=null
>> 2011-05-19 10:58:30,659 INFO
>> org.apache.hadoop.ipc.metrics.RpcMetrics: Initializing RPC Metrics
>> with hostName=DataNode, port=50020
>> 2011-05-19 10:58:30,670 INFO org.apache.hadoop.ipc.Server: IPC Server
>> Responder: starting
>> 2011-05-19 10:58:30,672 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 0 on 50020: starting
>> 2011-05-19 10:58:30,672 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: dnRegistration =
>> DatanodeRegistration(test.puppet.com:50010,
>> storageID=DS-173493047-127.0.0.1-50010-1305278767521, infoPort=50075,
>> ipcPort=50020)
>> 2011-05-19 10:58:30,673 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 1 on 50020: starting
>> 2011-05-19 10:58:30,689 INFO org.apache.hadoop.ipc.Server: IPC Server
>> handler 2 on 50020: starting
>> 2011-05-19 10:58:30,689 INFO org.apache.hadoop.ipc.Server: IPC Server
>> listener on 50020: starting
>> 2011-05-19 10:58:30,690 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode:
>> DatanodeRegistration(127.0.0.1:50010,
>> storageID=DS-173493047-127.0.0.1-50010-1305278767521, infoPort=50075,
>> ipcPort=50020)In DataNode.run, data =
>> FSDataset{dirpath='/tmp/hadoop-liyun2010/dfs/data/current'}
>> 2011-05-19 10:58:30,691 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: using
>> BLOCKREPORT_INTERVAL of 3600000msec Initial delay: 0msec
>> 2011-05-19 10:58:30,774 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0
>> blocks got processed in 23 msecs
>> 2011-05-19 10:58:30,776 INFO
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Starting Periodic
>> block scanner
>> 2011-05-19
>> ------------------------------------------------------------------------
>> 李�S
> Why don't you change the dfs dir from /tmp to another directory, for
> example /usr/share/hadoop/dfs?
> Can you attach your configuration files to inspect them?
>
> Regards
>
> --
> Marcos Luís Ortíz Valmaseda
> Software Engineer (Large-Scaled Distributed Systems)
> University of Information Sciences,
> La Habana, Cuba
> Linux User # 418229
> http://about.me/marcosortiz
--
Marcos Luís Ortíz Valmaseda
Software Engineer (Large-Scaled Distributed Systems)
University of Information Sciences,
La Habana, Cuba
Linux User # 418229
http://about.me/marcosortiz