Hi Huang,
Just run following command to for which application port 50011 is used.
netstat -anlp | grep 50011 
You will get process name and process id with above command, by using process 
id just kill the application running on that port.After killing try to start 
datanode.This solution works in Linux based operating systems.
Thanks,Sandeep.

From: [email protected]
Date: Wed, 26 Jun 2013 15:32:45 +0530
Subject: Re: datanode can not start
To: [email protected]; [email protected]

Remove 

<property>
       <name>dfs.datanode.address</name>
       <value>0.0.0.0:50011</value>



</property>


And try.





Thanks & Regards            


        
        
        
        


 ∞

Shashwat Shriparv



On Wed, Jun 26, 2013 at 3:29 PM, varun kumar <[email protected]> wrote:


HI huang,




Some other service is running on the port or you did not stop the datanode 
service properly.




Regards,Varun Kumar.P






On Wed, Jun 26, 2013 at 3:13 PM, ch huang <[email protected]> wrote:

i have running old cluster datanode,so it exist some conflict, i changed 
default port, here is my hdfs-site.xml





 

<configuration>
       <property>
                <name>dfs.name.dir</name>
                <value>/data/hadoopnamespace</value>
        </property>
        <property>
                <name>dfs.data.dir</name>
                <value>/data/hadoopdata</value>
        </property>
        <property>
                <name>dfs.datanode.address</name>
                <value>0.0.0.0:50011</value>






        </property>
        <property>
                <name>dfs.permissions</name>
                <value>false</value>
        </property>
        <property>
                <name>dfs.datanode.max.xcievers</name>
                <value>4096</value>
        </property>
        <property>
                <name>dfs.webhdfs.enabled</name>
                <value>true</value>
        </property>
        <property>
                <name>dfs.http.address</name>
                <value>192.168.10.22:50070</value>






        </property>
</configuration>
 
 
2013-06-26 17:37:24,923 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting DataNode
STARTUP_MSG:   host = CH34/192.168.10.34






STARTUP_MSG:   args = []
STARTUP_MSG:   version = 0.20.2-cdh3u4
STARTUP_MSG:   build = file:///data/1/tmp/topdir/BUILD/hadoop-0.20.2-cdh3u4 -r 
214dd731e3bdb687cb55988d3f47dd9e248c5690; compiled by 'root' on Mon May  7 
14:03:02 PDT 2012






************************************************************/
2013-06-26 17:37:25,335 INFO org.apache.hadoop.security.UserGroupInformation: 
JAAS Configuration already set up for Hadoop, not re-installing.
2013-06-26 17:37:25,421 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Registered FSDatasetStatusMBean






2013-06-26 17:37:25,429 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Opened streaming server at 50011
2013-06-26 17:37:25,430 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Balancing bandwith is 1048576 bytes/s






2013-06-26 17:37:25,470 INFO org.mortbay.log: Logging to 
org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog
2013-06-26 17:37:25,513 INFO org.apache.hadoop.http.HttpServer: Added global 
filtersafety (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)






2013-06-26 17:37:25,518 INFO org.apache.hadoop.http.HttpServer: Port returned 
by webServer.getConnectors()[0].getLocalPort() before open() is -1. Opening the 
listener on 50075
2013-06-26 17:37:25,519 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
Waiting for threadgroup to exit, active threads is 0






2013-06-26 17:37:25,619 INFO 
org.apache.hadoop.hdfs.server.datanode.FSDatasetAsyncDiskService: Shutting down 
all async disk service threads...
2013-06-26 17:37:25,619 INFO 
org.apache.hadoop.hdfs.server.datanode.FSDatasetAsyncDiskService: All async 
disk service threads have been shut down.






2013-06-26 17:37:25,620 ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: 
java.net.BindException: Address already in use
        at sun.nio.ch.Net.bind(Native Method)
        at 
sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:124)






        at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:59)
        at 
org.mortbay.jetty.nio.SelectChannelConnector.open(SelectChannelConnector.java:216)
        at org.apache.hadoop.http.HttpServer.start(HttpServer.java:564)






        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:505)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:303)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1643)






        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1583)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1601)
        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1727)






        at 
org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1744)
2013-06-26 17:37:25,622 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down DataNode at CH34/192.168.10.34






************************************************************/



-- 
Regards,
Varun Kumar.P


                                          

Reply via email to