Made that change . Still the same error.

And why should fs.default.name set to file:/// ?  I am not running in 
pseudo-distributed mode. I am having two systems one is master and the other is 
slave.

Vishnu Viswanath

> On 25-Dec-2013, at 19:35, kishore alajangi <[email protected]> wrote:
> 
> Replace hdfs:// to file:/// in fs.default.name property.
> 
> 
>> On Wed, Dec 25, 2013 at 7:01 PM, Vishnu Viswanath 
>> <[email protected]> wrote:
>> Hi,
>> 
>> I am getting this error while starting the datanode in my slave system.
>> 
>> I read the JIRA HDFS-2515, it says it is because hadoop is using wrong conf 
>> file. 
>> 
>> 13/12/24 15:57:14 INFO impl.MetricsConfig: loaded properties from 
>> hadoop-metrics2.properties
>> 13/12/24 15:57:14 INFO impl.MetricsSourceAdapter: MBean for source 
>> MetricsSystem,sub=Stats registered.
>> 13/12/24 15:57:14 INFO impl.MetricsSystemImpl: Scheduled snapshot period at 
>> 10 second(s).
>> 13/12/24 15:57:14 INFO impl.MetricsSystemImpl: DataNode metrics system 
>> started
>> 13/12/24 15:57:15 INFO impl.MetricsSourceAdapter: MBean for source ugi 
>> registered.
>> 13/12/24 15:57:15 WARN impl.MetricsSystemImpl: Source name ugi already 
>> exists!
>> 13/12/24 15:57:15 ERROR datanode.DataNode: 
>> java.lang.IllegalArgumentException: Does not contain a valid host:port 
>> authority: file:///
>>     at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164)
>>     at 
>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:212)
>>     at 
>> org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:244)
>>     at 
>> org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:236)
>>     at 
>> org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:359)
>>     at 
>> org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:321)
>>     at 
>> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1712)
>>     at 
>> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1651)
>>     at 
>> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1669)
>>     at 
>> org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1795)
>>     at 
>> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1812)
>> 
>> But how do i check which conf file hadoop is using? or how do i set it?
>> 
>> These are my configurations:
>> 
>> core-site.xml
>> ------------------
>> <configuration>
>>     <property>
>>         <name>fs.defualt.name</name>
>>         <value>hdfs://master:9000</value>
>>     </property>
>> 
>>     <property>
>>         <name>hadoop.tmp.dir</name>
>>         <value>/home/vishnu/hadoop-tmp</value>
>>     </property>
>> </configuration>
>> 
>> hdfs-site.xml
>> --------------------
>> <configuration>
>>     <property>
>>         <name>dfs.replication</name>
>>         <value>2</value>
>>     </property>
>> </configuration>
>> 
>> mared-site.xml
>> --------------------
>> <configuration>
>>     <property>
>>         <name>mapred.job.tracker</name>
>>         <value>master:9001</value>
>>     </property>
>> </configuration>
>> 
>> any help,
> 
> 
> 
> -- 
> Thanks,
> Kishore.

Reply via email to