Indeed, this was the right answer, and in the morning the file system was as fresh as in the evening. Somebody already told me to move out of /tmp, but I didn't believe him then. sorry. Mark
On Tue, Feb 17, 2009 at 7:57 AM, Rasit OZDAS <[email protected]> wrote: > I agree with Amandeep, and results will remain forever, unless you manually > delete them. > > If we are on the right road, > change hadoop.tmp.dir property to be outside of /tmp, or changing > dfs.name.dir and dfs.data.dir should be enough for basic use (I didn't have > to change anything else). > > Cheers, > Rasit > > 2009/2/17 Amandeep Khurana <[email protected]> > > > Where are your namenode and datanode storing the data? By default, it > goes > > into the /tmp directory. You might want to move that out of there. > > > > Amandeep > > > > > > Amandeep Khurana > > Computer Science Graduate Student > > University of California, Santa Cruz > > > > > > On Mon, Feb 16, 2009 at 8:11 PM, Mark Kerzner <[email protected]> > > wrote: > > > > > Hi all, > > > > > > I consistently have this problem that I can run HDFS and restart it > after > > > short breaks of a few hours, but the next day I always have to reformat > > > HDFS > > > before the daemons begin to work. > > > > > > Is that normal? Maybe this is treated as temporary data, and the > results > > > need to be copied out of HDFS and not stored for long periods of time? > I > > > verified that the files in /tmp related to hadoop are seemingly intact. > > > > > > Thank you, > > > Mark > > > > > > > > > -- > M. Raşit ÖZDAŞ >
