[ 
https://issues.apache.org/jira/browse/NUTCH-2921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17463232#comment-17463232
 ] 

Sebastian Nagel commented on NUTCH-2921:
----------------------------------------

Hi [~markus17], yes, that makes sense. The configurations or requirements may 
change over time, and then the CrawlDb entries need to be adjusted.

- instead of adding a second numeric property - what about a boolean property 
{{scoring.depth.max.override}} or {{scoring.depth.max.enforce}}? I mean, you 
want to change the default and then apply it to all existing CrawlDb records, 
including those with a custom depth set during injection.
- as always: if a property is introduced, it should be documented in 
nutch-default.xml
- why is the depth adjusted only for orphaned records? Do I miss something?
- the patch does not apply because it includes changes (license, override 
annotations) which are already in the master branch. Could you rebase the patch 
to the current master?


> DepthScoringFilter option to reset max_depth
> --------------------------------------------
>
>                 Key: NUTCH-2921
>                 URL: https://issues.apache.org/jira/browse/NUTCH-2921
>             Project: Nutch
>          Issue Type: Improvement
>            Reporter: Markus Jelsma
>            Priority: Minor
>         Attachments: NUTCH-2921.patch
>
>
> Once a max_depth has been set, it cannot be unset or reset. Here's a dirty 
> solution to globally reset the maximum depth for all records.



--
This message was sent by Atlassian Jira
(v8.20.1#820001)

Reply via email to