[
https://issues.apache.org/jira/browse/NUTCH-2921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17463232#comment-17463232
]
Sebastian Nagel commented on NUTCH-2921:
----------------------------------------
Hi [~markus17], yes, that makes sense. The configurations or requirements may
change over time, and then the CrawlDb entries need to be adjusted.
- instead of adding a second numeric property - what about a boolean property
{{scoring.depth.max.override}} or {{scoring.depth.max.enforce}}? I mean, you
want to change the default and then apply it to all existing CrawlDb records,
including those with a custom depth set during injection.
- as always: if a property is introduced, it should be documented in
nutch-default.xml
- why is the depth adjusted only for orphaned records? Do I miss something?
- the patch does not apply because it includes changes (license, override
annotations) which are already in the master branch. Could you rebase the patch
to the current master?
> DepthScoringFilter option to reset max_depth
> --------------------------------------------
>
> Key: NUTCH-2921
> URL: https://issues.apache.org/jira/browse/NUTCH-2921
> Project: Nutch
> Issue Type: Improvement
> Reporter: Markus Jelsma
> Priority: Minor
> Attachments: NUTCH-2921.patch
>
>
> Once a max_depth has been set, it cannot be unset or reset. Here's a dirty
> solution to globally reset the maximum depth for all records.
--
This message was sent by Atlassian Jira
(v8.20.1#820001)