Hi Eason, Thanks a lot for your reply. But I do have another folder which in not inside /tmp. I did not use default settings. To make it clear, I will describe what happened: 1. hadoop namenode -format 2. start-all.sh 3. running fine, http://localhost:50070 is accessible 4. stop-all.sh 5. start-all.sh, http://localhost:50070 is NOT accessible Unless I format the namenode, the HDFS master http://localhost:50070/dfshealth.jsp is not accessible. So, I have to redo step 1, 2 again to gain access to http://localhost:50070/dfshealth.jsp. But all data would be lost after format.
William On Mon, Mar 8, 2010 at 1:02 AM, Eason.Lee <[email protected]> wrote: > 2010/3/8 William Kang <[email protected]> > > > Hi guys, > > Thanks for your replies. I did not put anything in /tmp. It's just that > > > > default setting of dfs.name.dir/dfs.data.dir is set to the subdir in /tmp > > every time when I restart the hadoop, the localhost:50070 does not show up. > > The localhost:50030 is fine. Unless I reformat namenode, I wont be able > to > > see the HDFS' web page at 50070. It did not clean /tmp automatically. But > > > > It's not you clean the /tmp dir. Some operation clean it automatically~~ > > > > after format, everything is gone, well, it is a format. I did not really > > see > > anything in log. Not sure what caused it. > > > > > > William > > > > > > On Mon, Mar 8, 2010 at 12:39 AM, Bradford Stephens < > > [email protected]> wrote: > > > > > Yeah. Don't put things in /tmp. That's unpleasant in the long run. > > > > > > On Sun, Mar 7, 2010 at 9:36 PM, Eason.Lee <[email protected]> wrote: > > > > Your /tmp directory is cleaned automaticly? > > > > > > > > Try to set dfs.name.dir/dfs.data.dir to a safe dir~~ > > > > > > > > 2010/3/8 William Kang <[email protected]> > > > > > > > >> Hi all, > > > >> I am running HDFS in Pseudo-distributed mode. Every time after I > > > restarted > > > >> the machine, I have to format the namenode otherwise the > > localhost:50070 > > > >> wont show up. It is quite annoying to do so since all the data would > > be > > > >> lost. Does anybody know this happens? And how should I fix this > > problem? > > > >> Many thanks. > > > >> > > > >> > > > >> William > > > >> > > > > > > > > > > > > > > > > -- > > > http://www.drawntoscalehq.com -- The intuitive, cloud-scale data > > > solution. Process, store, query, search, and serve all your data. > > > > > > http://www.roadtofailure.com -- The Fringes of Scalability, Social > > > Media, and Computer Science > > > > > >
