[ 
https://issues.apache.org/jira/browse/NUTCH-2328?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15586690#comment-15586690
 ] 

Sebastian Nagel commented on NUTCH-2328:
----------------------------------------

> the only solution is to have a cluster wide propagated count 

No, this is not required. The solution with an instance variable is by design:
- local, per-reducer limit = topN / number of reducers
- every reducer checks only for the local limit
- in sum, there will be topN URLs generated

The condition is that URLs are evenly distributed across different hosts (at 
least as many as there are reducers), cf. 
[[1|https://www.mail-archive.com/user@nutch.apache.org/msg14499.html]].

A job-wide counter does not guarantee any limits betters because there is no 
control how reduce tasks are launched in time. Only if all tasks run in 
parallel, with similar speed and no task fails, an even distribution across 
reducers/parts would be achieved. But that will hardly happen in a production 
Hadoop cluster. In a realistic scenario some tasks are launched first and will 
get more URLs. The tasks launched later get less or even no URLs. However, to 
achieve an optimal utilization of the fetcher, all parts should be of equal 
size.

> GeneratorJob does not generate anything on second run
> -----------------------------------------------------
>
>                 Key: NUTCH-2328
>                 URL: https://issues.apache.org/jira/browse/NUTCH-2328
>             Project: Nutch
>          Issue Type: Bug
>          Components: generator
>    Affects Versions: 2.2, 2.3, 2.2.1, 2.3.1
>         Environment: Ubuntu 16.04 / Hadoop 2.7.1
>            Reporter: Arthur B
>              Labels: fails, generator, subsequent
>             Fix For: 2.4
>
>         Attachments: generator-issue-static-count.patch
>
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> Given a topN parameter (ie 10) the GeneratorJob will fail to generate 
> anything new on the subsequent runs within the same process space.
> To reproduce the issue submit the GeneratorJob twice one after another to the 
> M/R framework. Second time will say it generated 0 URLs.
> This issue is due to the usage of the static count field 
> (org.apache.nutch.crawl.GeneratorReducer#count) to determine if the topN 
> value has been reached.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to