There may be no relevant links if all documents are from one single host (or domain) and (link.ignore.internal.host == true) resp. (link.ignore.internal.domain == true) cf. explanations about that in the wiki.
2014-03-26 4:09 GMT+01:00 John Lafitte <[email protected]>: > Thanks for that Sebastian. So given the hint you've given me, I'm trying > to generate the scoring using this example: > https://wiki.apache.org/nutch/NewScoringIndexingExample > > But when it gets to the LinkRank part I get: > > 2014-03-26 02:57:14,208 INFO webgraph.LinkRank - Analysis: starting at > 2014-03-26 02:57:14 > 2014-03-26 02:57:14,913 INFO webgraph.LinkRank - Starting link counter job > 2014-03-26 02:57:17,927 INFO webgraph.LinkRank - Finished link counter job > 2014-03-26 02:57:17,928 INFO webgraph.LinkRank - Reading numlinks temp > file > 2014-03-26 02:57:17,932 ERROR webgraph.LinkRank - LinkAnalysis: > java.io.IOException: No links to process, is the webgra$ > at > org.apache.nutch.scoring.webgraph.LinkRank.runCounter(LinkRank.java:132) > at > org.apache.nutch.scoring.webgraph.LinkRank.analyze(LinkRank.java:622) > at > org.apache.nutch.scoring.webgraph.LinkRank.run(LinkRank.java:702) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > at > org.apache.nutch.scoring.webgraph.LinkRank.main(LinkRank.java:668) > > I can see the webgraph directory got created and there are directories and > files in there, but I'm guessing something is not getting populated > correctly. Any clue what I may be doing wrong? > > > On Tue, Mar 25, 2014 at 4:15 PM, Sebastian Nagel < > [email protected] > > wrote: > > > Hi John, > > > > FreeGenerator unlike Injector does not use db.score.injected (default = > > 1.0) > > but sets the initial score to 0.0. If all URLs stem from FreeGenerator > the > > total > > score in the link graph is also 0.0, and no linked documents can get a > > higher score > > that 0.0 > > As possible solutions: > > - use FreeGenerator with a initial score > 0.0 > > (but don't put thousands URLs with a score of 1.0: > > if the total score is too high some pages may get unreasonable > > high scores) > > - use linkrank (https://wiki.apache.org/nutch/NewScoring) to get the > > scores: > > the default scoring OPIC has the advantage of calculating scores online > > while following links. It gives good and plausible scores if crawl is > > started > > from few authoritative seeds. But sometimes, esp. in continuous crawls, > > OPIC scores run out of control. > > > > Sebastian > > > > On 03/25/2014 08:31 PM, John Lafitte wrote: > > > I setup a script that uses freegen to manually index new/updated URLs. > I > > > thought it was working great, but now I'm just realizing that Solr > > returns > > > a score of 0 for these new documents. I thought the score was > calculated > > > independent from what Nutch does, just uses the content and other > > metadata > > > to calculate it, however that doesn't seem to be the case. Anyone > have a > > > clue what might be causing this? The content and other metadata look > > > normal and I reloaded the core to no avail. > > > > > > > >

