Shashang Sheth created HDFS-6725: ------------------------------------ Summary: Datanode and Task Tracker not starting Key: HDFS-6725 URL: https://issues.apache.org/jira/browse/HDFS-6725 Project: Hadoop HDFS Issue Type: Bug Components: datanode Environment: user@Hadslave1:~$ uname -a Linux Hadslave1 3.2.0-23-generic-pae #36-Ubuntu SMP Tue Apr 10 22:19:09 UTC 2012 i686 i686 i386 GNU/Linux user@Hadslave1:~$
user@Hadslave1:~$ /opt/hadoop-1.2.0/bin/hadoop -version java version "1.6.0_31" OpenJDK Runtime Environment (IcedTea6 1.13.3) (6b31-1.13.3-1ubuntu1~0.12.04.2) OpenJDK Client VM (build 23.25-b01, mixed mode, sharing) user@Hadslave1:~$ java -version java version "1.6.0_31" OpenJDK Runtime Environment (IcedTea6 1.13.3) (6b31-1.13.3-1ubuntu1~0.12.04.2) OpenJDK Client VM (build 23.25-b01, mixed mode, sharing) user@Hadslave1:~$ Reporter: Shashang Sheth Priority: Minor I am unable to start the datanode and tasktracker daemons on one of my slave nodes. I have got two slave nodes for my test env. The error is familiar to Jira, however the solutions provided in JIRA is not working for me. Below are the errors Datanode log file: 2014-07-22 07:17:54,559 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: STARTUP_MSG: /************************************************************ STARTUP_MSG: Starting DataNode STARTUP_MSG: host = Hadslave1/127.0.1.1 STARTUP_MSG: args = [] STARTUP_MSG: version = 1.2.0 STARTUP_MSG: build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.2 -r 1479473; compiled by 'hortonfo' on Mon May 6 06:59:37 UTC 2013 STARTUP_MSG: java = 1.6.0_31 ************************************************************/ 2014-07-22 07:17:55,691 INFO org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from hadoop-metrics2.properties 2014-07-22 07:17:55,703 INFO org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source MetricsSystem,sub=Stats registered. 2014-07-22 07:17:55,732 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s). 2014-07-22 07:17:55,732 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: DataNode metrics system started 2014-07-22 07:17:56,265 INFO org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi registered. 2014-07-22 07:17:56,275 WARN org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name ugi already exists! 2014-07-22 07:17:57,536 ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: java.lang.IllegalArgumentException: Does not contain a valid host:port authority: file:/// at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:212) at org.apache.hadoop.hdfs.server.namenode.NameNode.getAddress(NameNode.java:244) at org.apache.hadoop.hdfs.server.namenode.NameNode.getServiceAddress(NameNode.java:236) at org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:357) at org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:319) at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1698) at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1637) at org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1655) at org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1781) at org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1798) 2014-07-22 07:17:57,566 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down DataNode at Hadslave1/127.0.1.1 ************************************************************/ Task tracker's log files: 2014-07-22 07:17:59,297 INFO org.apache.hadoop.mapred.TaskTracker: STARTUP_MSG: /************************************************************ STARTUP_MSG: Starting TaskTracker STARTUP_MSG: host = Hadslave1/127.0.1.1 STARTUP_MSG: args = [] STARTUP_MSG: version = 1.2.0 STARTUP_MSG: build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.2 -r 1479473; compiled by 'hortonfo' on Mon May 6 06:59:37 UTC 2013 STARTUP_MSG: java = 1.6.0_31 ************************************************************/ 2014-07-22 07:17:59,671 INFO org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from hadoop-metrics2.properties 2014-07-22 07:17:59,814 INFO org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source MetricsSystem,sub=Stats registered. 2014-07-22 07:17:59,815 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s). 2014-07-22 07:17:59,815 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: TaskTracker metrics system started 2014-07-22 07:18:00,028 INFO org.apache.hadoop.util.NativeCodeLoader: Loaded the native-hadoop library 2014-07-22 07:18:00,158 INFO org.apache.hadoop.metrics2.impl.MetricsSourceAdapter: MBean for source ugi registered. 2014-07-22 07:18:00,160 WARN org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Source name ugi already exists! 2014-07-22 07:18:00,265 ERROR org.apache.hadoop.mapred.TaskTracker: Can not start task tracker because java.lang.IllegalArgumentException: Does not contain a valid host:port authority: local at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:164) at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) at org.apache.hadoop.mapred.JobTracker.getAddress(JobTracker.java:2121) at org.apache.hadoop.mapred.TaskTracker.<init>(TaskTracker.java:1540) at org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:3937) 2014-07-22 07:18:00,265 INFO org.apache.hadoop.mapred.TaskTracker: SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down TaskTracker at Hadslave1/127.0.1.1 ************************************************************/ Below are my configuration files: user@Hadmast:/opt/hadoop-1.2.0/conf$ cat core-site.xml <?xml version="1.0"?> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> <!-- Put site-specific property overrides in this file. --> <configuration> <property> <name>fs.default.name</name> <value>hdfs://192.168.111.131:8020</value> </property> </configuration> user@Hadmast:/opt/hadoop-1.2.0/conf$ cat hdfs-site.xml <?xml version="1.0"?> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> <!-- Put site-specific property overrides in this file. --> <configuration> <property> <name>dfs.replication</name> <value>2</value> </property> <property> <name>dfs.permissions</name> <value>false</value> </property> </configuration> user@Hadmast:/opt/hadoop-1.2.0/conf$ cat mapred-site.xml <?xml version="1.0"?> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> <!-- Put site-specific property overrides in this file. --> <configuration> <property> <name>mapred.job.tracker</name> <value>localhost:8021</value> </property> </configuration> user@Hadmast:/opt/hadoop-1.2.0/conf$ Both the daemons start on Hadmast node but doesn't start on the other node. user@Hadmast:/opt/hadoop-1.2.0/conf$ jps 6421 NameNode 7592 Jps 6941 JobTracker 6866 SecondaryNameNode 7172 TaskTracker user@Hadslave1:~$ jps 4826 Jps user@Hadslave1:~$ I have formated the namenode multiple times and have also rebuilt the Hadslave1 once but no change. -- This message was sent by Atlassian JIRA (v6.2#6252)