Thank you all for the responses.

So in order to run a web-based application, I just need to put the part of
the application that needs to make use of distributed computation in HDFS,
and have the other web site related files access it via Hadoop streaming ?

Is that how Hadoop is used ?

Sorry the question may sound too silly.

Thank you.


On Thu, Jun 12, 2008 at 7:49 PM, Ted Dunning <[EMAIL PROTECTED]> wrote:

> Once it is in HDFS, you already have backups (due to the replicated file
> system).
>
> Your problems with deleting the dfs data directory are likely configuration
> problems combined with versioning of the data store (done to avoid
> confusion, but usually causes confusion).  Once you get the configuration
> and operational issues sorted out, you shouldn't lose any data.
>
> On Thu, Jun 12, 2008 at 10:15 AM, Chanchal James <[EMAIL PROTECTED]>
> wrote:
>
> >
> > If I keep all data in HDFS, is there anyway I can back it up regularly.
> >
> >
>

Reply via email to