> 2008-09-05 10:11:20,106 ERROR org.apache.hadoop.dfs.DataNode:
> java.net.BindException: Problem binding to /0.0.0.0:50010

Should you check if something is listening on 50010? I strongly
encourage users to go through these logs when something does not work.

Regd running multiple datanodes :

Along with config directory, you should change the following env
variables : HADOOP_LOG_DIR and HADOOP_PID_DIR (I set both to the same
value).

The following variables should be different in hadoop-site.xml :
1. dfs.data.dir or hadoop.tmp.dir
2. dfs.datanode.address
3. dfs.datanode.http.address
4. dfs.datanode.ipc.address

Raghu.

叶双明 wrote:
> In addition, I can't start datanode in another computer use command:
> bin/hadoop-daemons.sh --config conf start datanode, by the default config,
> log message is:
> 
> 2008-09-05 10:11:19,208 INFO org.apache.hadoop.dfs.DataNode: STARTUP_MSG:
> /************************************************************
> STARTUP_MSG: Starting DataNode
> STARTUP_MSG:   host = testcenter/192.168.100.120
> STARTUP_MSG:   args = []
> STARTUP_MSG:   version = 0.17.1
> STARTUP_MSG:   build =
> http://svn.apache.org/repos/asf/hadoop/core/branches/branch-0.17 -r 669344;
> compiled by 'hadoopqa' on Thu Jun 19 01:18:25 UTC 2008
> ************************************************************/
> 2008-09-05 10:11:20,097 INFO org.apache.hadoop.dfs.DataNode: Registered
> FSDatasetStatusMBean
> 2008-09-05 10:11:20,106 ERROR org.apache.hadoop.dfs.DataNode:
> java.net.BindException: Problem binding to /0.0.0.0:50010
>         at org.apache.hadoop.ipc.Server.bind(Server.java:175)
>         at org.apache.hadoop.dfs.DataNode.startDataNode(DataNode.java:264)
>         at org.apache.hadoop.dfs.DataNode.<init>(DataNode.java:171)
>         at org.apache.hadoop.dfs.DataNode.makeInstance(DataNode.java:2765)
>         at
> org.apache.hadoop.dfs.DataNode.instantiateDataNode(DataNode.java:2720)
>         at org.apache.hadoop.dfs.DataNode.createDataNode(DataNode.java:2728)
>         at org.apache.hadoop.dfs.DataNode.main(DataNode.java:2850)
> 
> 2008-09-05 10:11:20,107 INFO org.apache.hadoop.dfs.DataNode: SHUTDOWN_MSG:
> /************************************************************
> SHUTDOWN_MSG: Shutting down DataNode at testcenter/192.168.100.120
> ************************************************************/
> 
> Maybe there is some wrong network config???
> 
> 2008/9/5 叶双明 <[EMAIL PROTECTED]>
> 
>> I can start first datanode by
>> bin/hadoop-daemons.sh --config conf1 start datanode
>>
>> and copy conf1 to conf2, and modify the value for follwing properties, port
>> = port +1
>>
>> <property>
>>   <name>dfs.datanode.address</name>
>>   <value>0.0.0.0:50011</value>
>>   <description>
>>     The address where the datanode server will listen to.
>>     If the port is 0 then the server will start on a free port.
>>   </description>
>> </property>
>>
>> <property>
>>   <name>dfs.datanode.http.address</name>
>>   <value>0.0.0.0:50076</value>
>>   <description>
>>     The datanode http server address and port.
>>     If the port is 0 then the server will start on a free port.
>>   </description>
>> </property>
>>
>> <property>
>>   <name>dfs.http.address</name>
>>   <value>0.0.0.0:50071</value>
>>   <description>
>>     The address and the base port where the dfs namenode web ui will listen
>> on.
>>     If the port is 0 then the server will start on a free port.
>>   </description>
>> </property>
>>
>> <property>
>>   <name>dfs.datanode.https.address</name>
>>   <value>0.0.0.0:50476</value>
>> </property>
>>
>> And run bin/hadoop-daemons.sh --config conf2 start datanode
>> but get message: datanode running as process 4137. Stop it first.
>>
>> what other properties could I modify between the two config file.
>>
>> Thanks.
>>
>> 2008/9/5 lohit <[EMAIL PROTECTED]>
>>
>> Good way is to have different 'conf' dirs
>>> So, you would end up with dir conf1 and conf2
>>> and startup of datanode would be
>>> ./bin/hadoop-daemons.sh --config conf1 start datanode
>>> ./bin/hadoop-daemons.sh --config conf2 start datanode
>>>
>>>
>>> make sure you have different hadoop-site.xml in conf1 and conf2 dirs.
>>>
>>> -Lohit
>>>
>>> ----- Original Message ----
>>> From: 叶双明 <[EMAIL PROTECTED]>
>>> To: core-user@hadoop.apache.org
>>> Sent: Thursday, September 4, 2008 12:01:48 AM
>>> Subject: Re: can i run multiple datanode in one pc?
>>>
>>> Thanks lohit.
>>>
>>> I run start datanod by comman: bin/hadoop datanode -conf
>>> conf/hadoop-site.xml, it can't work, but command: bin/hadoop datanode can
>>> work.
>>>
>>> Something wrong have I done?
>>>
>>>
>>>
>>> bin/hadoop datanode -conf conf/hadoop-site.xml
>>>
>>>
>>> 2008/9/4 lohit <[EMAIL PROTECTED]>
>>>
>>>> Yes, each datanode should point to different config.
>>>> So, if you have conf/hadoop-site.xml make another conf2/hadoop-site.xml
>>>> with ports for datanode specific stuff and you should be able to start
>>>> multiple datanodes on same node.
>>>> -Lohit
>>>>
>>>>
>>>>
>>>> ----- Original Message ----
>>>> From: 叶双明 <[EMAIL PROTECTED]>
>>>> To: core-user@hadoop.apache.org
>>>> Sent: Wednesday, September 3, 2008 8:19:59 PM
>>>> Subject: can i run multiple datanode in one pc?
>>>>
>>>> i know it is running one datanode in one computer normally。
>>>> i wondering can i run multiple datanode in one pc?
>>>>
>>>>
>>>

Reply via email to