yes it should if you configured hadoop correctly, just run the hadoop
balancer afterwards to balance the HDFS use.

$ bin/start-balancer.sh


Have a good day :)


On Fri, Oct 8, 2010 at 7:36 PM, MilleBii <[email protected]> wrote:

> for Q2 I'm already using HDFS but in pseudo distributed, ie I have hadoop
> running on a single server.
> So if I had a server it should run ? That would save a lot of time.
>
> -Raymond-
>
> 2010/10/8 CatOs Mandros <[email protected]>
>
> > Hi MilleBii,
> >
> > On Fri, Oct 8, 2010 at 7:21 PM, MilleBii <[email protected]> wrote:
> >
> > > How easy is it scale-up nutch following demand ?
> > >
> > >
> > Its kinda easy when you already have a distributed environement working.
> >
> >
> > > Q1 : Suppose I have 2 servers and I want to add a 3 third one. Modulo
> the
> > > adequate configuration will hadoop just add the server and manage to
> > > distribute the data on hdfs: or I need to do something special.
> > >
> >
> > You have to add the machine to the cluster and hadoop will pick it up.
> >
> > Just configure the machine,
> >
> > start the services:
> > $ cd path/to/hadoop
> > $ bin/hadoop-daemon.sh start datanode
> > $ bin/hadoop-daemon.sh start tasktracker
> >
> > and you should be up and running.
> >
> >
> >
> > >
> > > Q2 : other question, from single server/pseudo distributed to a 2
> server
> > > distributed. It is just a matter of configuring hadoop or is there more
> > to
> > > it ?
> > >
> > >
> > Just configuring hadoop and telling to nutch to use the distributed mode.
> > You'll have to migrate your data to HDFS tough.
> >
> > --
> > > -MilleBii-
> > >
> >
>
>
>
> --
> -MilleBii-
>

Reply via email to