Hello, everyone. I met a strange problem and I am looking forward to your help.
I setup a 2 nodes cluster hadoop on 2 mechine runing ubuntu. and 1 is both master and slave and the other one is slave, and it seems to be functional. but yesterday, it does not work anymore. when I type hadoop command such as "bin/hadoop dfs -ls", it shows that "Bad connection to FS. command aborted." so, I run "bin/start-all.sh" to see if the dfs is shut, however it shows that: (***.***.***.*** stands for the master's ip address; ###.###.### stands for the other one's ip address) starting namenode, logging to ***** ***.***.***.***: datanode running as process 14800. Stop it first. ***.***.***.***: datanode running as process 9113. Stop it first. ***.***.***.***: starting secondarynamenode, logging to *** jobtracker running as process 14946. Stop it first. ***.***.***.***: tasktracker running as process 15029. Stop it first. ###.###.###.###: tasktracker running as process 9217. Stop it first. and i type "bin/hadoop dfs -ls", but again: Bad connection to FS. command aborted. then when I type "bin/hadoop dfs -ls", it shows: (***.***.***.*** stands for the master's ip address; ###.###.### stands for the other one's ip address) stopping jobtracker ***.***.***.***: stopping tasktracker ###.###.###: stopping tasktracker no namenode to stop <--------this is very strange ***.***.***.***: stopping datanode ###.###.###: stopping datanode ***.***.***.***: no secondarynamenode to stop when i restart it by "bin/start-all.sh", it still does not work.. can anyone give me some idea? Thanks a lot. -- Yours Sincerely Xudong Du Zijing 2# 305A Tsinghua University, Beijing, China, 100084
