RE: hadoop start problem
I had a similar problem when I upgraded... not sure of details why, but I had permissions problems trying to develop and run on windows out of cygwin. I found that in cygwin if I ran under my account, I got the null pointer exception, but if I shh localhost first, then format the name node, Hadoop ran as SYSTEM and initialized and ran properly. Somehow your permissions are likely the problem.. -Original Message- From: Aaron Kimball [mailto:[EMAIL PROTECTED] Sent: Monday, November 10, 2008 4:39 AM To: core-user@hadoop.apache.org Subject: Re: hadoop start problem Between 0.15 and 0.18 the format for fs.default.name has changed; you should set the value there as hdfs://localhost:9000/ without the quotes. It still shouldn't give you a NPE (that should probably get a JIRA entry) under any circumstances, but putting a value in the (new) proper format might get your system working. - Aaron On Sat, Nov 8, 2008 at 1:59 AM, ZhiHong Fu [EMAIL PROTECTED] wrote: Hi: I have encountered a strange problem,. I installed hadoop 0.15.2 on my computer servral month ago. but now I want to upgrade to 0.18.1, I delete the directory hadoop-0.15.2 and copy the hadoop 0.18.1 , and did some very simple configuration following the hadoop pre-sudo distributed tutorial and modified the hadoop-site.xml : configuration property namefs.default.name/name valuelocalhost:9000/value /property property namemapred.job.tracker/name valuelocalhost:9001/value /property property namedfs.replication/name value1/value /property /configuration I can run the bin/hadoop namenode -format , but when i run the command bin/start-dfs.sh , It will throw the null pointer exception. localhost: Exception in thread main java.lang.NullPointerException localhost: at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) localhost: at org.apache.hadoop.dfs.NameNode.getAddress(NameNode.java:116) localhost: at org.apache.hadoop.dfs.NameNode.getAddress(NameNode.java:120) localhost: at org.apache.hadoop.dfs.*SecondaryNameNode*.initialize(*SecondaryNameNode* .java:124) localhost: at org.apache.hadoop.dfs.*SecondaryNameNode*.init(*SecondaryNameNode*.jav a:108) localhost: at org.apache.hadoop.dfs.*SecondaryNameNode*.main(*SecondaryNameNode*.java: 460) I don't know why, I can run hadoop-0.15.2 very well _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ The information transmitted is intended only for the person or entity to which it is addressed and may contain confidential and/or privileged material. Any review, retransmission, dissemination or other use of, or taking of any action in reliance upon, this information by persons or entities other than the intended recipient is prohibited. If you received this message in error, please contact the sender and delete the material from any computer.
Re: hadoop start problem
Between 0.15 and 0.18 the format for fs.default.name has changed; you should set the value there as hdfs://localhost:9000/ without the quotes. It still shouldn't give you a NPE (that should probably get a JIRA entry) under any circumstances, but putting a value in the (new) proper format might get your system working. - Aaron On Sat, Nov 8, 2008 at 1:59 AM, ZhiHong Fu [EMAIL PROTECTED] wrote: Hi: I have encountered a strange problem,. I installed hadoop 0.15.2 on my computer servral month ago. but now I want to upgrade to 0.18.1, I delete the directory hadoop-0.15.2 and copy the hadoop 0.18.1 , and did some very simple configuration following the hadoop pre-sudo distributed tutorial and modified the hadoop-site.xml : configuration property namefs.default.name/name valuelocalhost:9000/value /property property namemapred.job.tracker/name valuelocalhost:9001/value /property property namedfs.replication/name value1/value /property /configuration I can run the bin/hadoop namenode -format , but when i run the command bin/start-dfs.sh , It will throw the null pointer exception. localhost: Exception in thread main java.lang.NullPointerException localhost: at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) localhost: at org.apache.hadoop.dfs.NameNode.getAddress(NameNode.java:116) localhost: at org.apache.hadoop.dfs.NameNode.getAddress(NameNode.java:120) localhost: at org.apache.hadoop.dfs.*SecondaryNameNode*.initialize(*SecondaryNameNode*.java:124) localhost: at org.apache.hadoop.dfs.*SecondaryNameNode*.init(*SecondaryNameNode*.java:108) localhost: at org.apache.hadoop.dfs.*SecondaryNameNode*.main(*SecondaryNameNode*.java:460) I don't know why, I can run hadoop-0.15.2 very well
Re: hadoop start problem
On 11/10/08 6:18 AM, Brian MacKay [EMAIL PROTECTED] wrote: I had a similar problem when I upgraded... not sure of details why, but I had permissions problems trying to develop and run on windows out of cygwin. At Apachecon, we think we identified a case where someone forgot to copy the newer hadoop-defaults.xml into their old configuration directories that they were using post-upgrade. Hadoop acts really strangely under those conditions.
hadoop start problem
Hi: I have encountered a strange problem,. I installed hadoop 0.15.2 on my computer servral month ago. but now I want to upgrade to 0.18.1, I delete the directory hadoop-0.15.2 and copy the hadoop 0.18.1 , and did some very simple configuration following the hadoop pre-sudo distributed tutorial and modified the hadoop-site.xml : configuration property namefs.default.name/name valuelocalhost:9000/value /property property namemapred.job.tracker/name valuelocalhost:9001/value /property property namedfs.replication/name value1/value /property /configuration I can run the bin/hadoop namenode -format , but when i run the command bin/start-dfs.sh , It will throw the null pointer exception. localhost: Exception in thread main java.lang.NullPointerException localhost: at org.apache.hadoop.net.NetUtils.createSocketAddr(NetUtils.java:130) localhost: at org.apache.hadoop.dfs.NameNode.getAddress(NameNode.java:116) localhost: at org.apache.hadoop.dfs.NameNode.getAddress(NameNode.java:120) localhost: at org.apache.hadoop.dfs.*SecondaryNameNode*.initialize(*SecondaryNameNode*.java:124) localhost: at org.apache.hadoop.dfs.*SecondaryNameNode*.init(*SecondaryNameNode*.java:108) localhost: at org.apache.hadoop.dfs.*SecondaryNameNode*.main(*SecondaryNameNode*.java:460) I don't know why, I can run hadoop-0.15.2 very well