[ https://issues.apache.org/jira/browse/NUTCH-2921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17463232#comment-17463232 ]
Sebastian Nagel commented on NUTCH-2921: ---------------------------------------- Hi [~markus17], yes, that makes sense. The configurations or requirements may change over time, and then the CrawlDb entries need to be adjusted. - instead of adding a second numeric property - what about a boolean property {{scoring.depth.max.override}} or {{scoring.depth.max.enforce}}? I mean, you want to change the default and then apply it to all existing CrawlDb records, including those with a custom depth set during injection. - as always: if a property is introduced, it should be documented in nutch-default.xml - why is the depth adjusted only for orphaned records? Do I miss something? - the patch does not apply because it includes changes (license, override annotations) which are already in the master branch. Could you rebase the patch to the current master? > DepthScoringFilter option to reset max_depth > -------------------------------------------- > > Key: NUTCH-2921 > URL: https://issues.apache.org/jira/browse/NUTCH-2921 > Project: Nutch > Issue Type: Improvement > Reporter: Markus Jelsma > Priority: Minor > Attachments: NUTCH-2921.patch > > > Once a max_depth has been set, it cannot be unset or reset. Here's a dirty > solution to globally reset the maximum depth for all records. -- This message was sent by Atlassian Jira (v8.20.1#820001)