Hi Hannes,

I'm curious as to whether you got this configuration running, any issues you ran into, and what performance you saw.

Thanks,

-- Ken


On Nov 20, 2010, at 10:52am, Hannes Carl Meyer wrote:

Ken, thanks, I guess thats a good hint!

I'm using the simple org.apache.nutch.crawl.Crawl to perform the crawl - I
guess the configuration of the Map-Reduce Job then is pretty low.

@Andrzej could you give me a hint where to configure the number of reduce
tasks in nutch 0.9? (running on a single machine)

Regards,

Hannes

On Sat, Nov 20, 2010 at 7:06 PM, Ken Krugler <[email protected] >wrote:


On Nov 20, 2010, at 7:51am, Hannes Carl Meyer wrote:

Thank you for sharing your experiences!

in my case the web servers are pretty stable and we are allowed to perform intensive crawling which make it easy to increase the threads per host.

imho the fetch process isn't really the bottleneck. It is the process
between the fetch process when merging and updating the crawldb.

We are using a 16 Core Hardware, during fetch process CPUs are being used around 1000 % but in between fetching it is always around 90-100 % on a
single core


In regular map-reduce Hadoop jobs you get this situation if the job has been configured to use a single reducer, and thus only one core is active

Though it would surprise me if the crawlDB update job was configured this way, as I don't see a reason why the crawlDB has to be a single file in
HDFS.

Andrzej and others would know best, of course.

-- Ken




On Sat, Nov 20, 2010 at 11:33 AM, Ye T Thet <[email protected]>
wrote:

Hannes,

I guess It would depends on situation
- your server specs (where cralwer is running) and
- hosts specs

Anyway, I have been crawling around 50 hosts. I tweaked a few to get it
right for my situation.

Currently I am using 500 threads. and 10 threads per host.

In my opinion, number of threads for crawler does not matter much.
Because
crawler does not take much of a resource (memory and CPU). As far as your
server network band width can handle, it should be fine.

In my case, number of threads per host matters. Because some of my server
cannot handle that much of bandwidth.

Not sure if it would helps, I had to adjust fetcher.server.delay,
fetcher.server.min.delay and fetcher.max.crawl.delay because, my hosts
sometimes cannot handle that much of threads.


Warm Regards,

Y.T. Thet




On Thu, Nov 18, 2010 at 11:06 PM, Hannes Carl Meyer <
[email protected]> wrote:

Hi Ken,

our Crawler is allowed to hit those hosts in a frequent way at night so
we
are not getting a penalty ;-)

Could you imagine running nutch in this case with about 400 threads,
with
1
thread per host and a delay of 1.0?

I tried that way but experienced some really long idle times... My idea
was
one thread per host. That would mean adding another host would require
add
an additional thread.

Regards

Hannes

On Thu, Nov 18, 2010 at 3:36 PM, Ken Krugler <
[email protected]

wrote:


If you're hitting each host with 45 threads, you better be on really

good

terms with those webmasters :)

With 90 total threads, that means as few as 2 hosts are active at any

time,

yes?

-- Ken



On Nov 18, 2010, at 3:51am, Hannes Carl Meyer wrote:

Hi,

I'm using nutch 0.9 to crawl about 400 hosts with an average of 600

pages.

That makes a volume of 240.000 fetched pages - I want to get all of

them.


Can one give me an advice on the right threads/delay/per-host
configuration
in this environnement?

My current conf:

<property>
   <name>fetcher.server.delay</name>
   <value>1.0</value>
</property>

<property>
   <name>fetcher.threads.fetch</name>
   <value>90</value>
</property>

<property>
   <name>fetcher.threads.per.host</name>
   <value>45</value>
</property>

<property>
 <name>fetcher.threads.per.host.by.ip</name>
 <value>false</value>
</property>

The total runtime is about 5 hours.

How can performance be improved? (I still have enough CPU, Bandwith)

Note: This runs on a single machine, distribution to other machines is

not

planned.

Thanks and Regards

Hannes


--------------------------
Ken Krugler
+1 530-210-6378
http://bixolabs.com
e l a s t i c   w e b   m i n i n g










--------------------------
Ken Krugler
+1 530-210-6378
http://bixolabs.com
e l a s t i c   w e b   m i n i n g







--------------------------
Ken Krugler
+1 530-210-6378
http://bixolabs.com
e l a s t i c   w e b   m i n i n g





Reply via email to