Indeed, this was the right answer, and in the morning the file system was as
fresh as in the evening. Somebody already told me to move out of /tmp, but I
didn't believe him then. sorry.
Mark

On Tue, Feb 17, 2009 at 7:57 AM, Rasit OZDAS <[email protected]> wrote:

> I agree with Amandeep, and results will remain forever, unless you manually
> delete them.
>
> If we are on the right road,
> change hadoop.tmp.dir property to be outside of /tmp, or changing
> dfs.name.dir and dfs.data.dir should be enough for basic use (I didn't have
> to change anything else).
>
> Cheers,
> Rasit
>
> 2009/2/17 Amandeep Khurana <[email protected]>
>
> > Where are your namenode and datanode storing the data? By default, it
> goes
> > into the /tmp directory. You might want to move that out of there.
> >
> > Amandeep
> >
> >
> > Amandeep Khurana
> > Computer Science Graduate Student
> > University of California, Santa Cruz
> >
> >
> > On Mon, Feb 16, 2009 at 8:11 PM, Mark Kerzner <[email protected]>
> > wrote:
> >
> > > Hi all,
> > >
> > > I consistently have this problem that I can run HDFS and restart it
> after
> > > short breaks of a few hours, but the next day I always have to reformat
> > > HDFS
> > > before the daemons begin to work.
> > >
> > > Is that normal? Maybe this is treated as temporary data, and the
> results
> > > need to be copied out of HDFS and not stored for long periods of time?
> I
> > > verified that the files in /tmp related to hadoop are seemingly intact.
> > >
> > > Thank you,
> > > Mark
> > >
> >
>
>
>
> --
> M. Raşit ÖZDAŞ
>

Reply via email to