Hi Thomas, After adding the properties as you mentioned, did you re-start at the injecting stage or did you just use the webgraph class? If the latter, then I would try re-starting the whole process, maybe even checking you reading your crawldb on the way to executing the webgraph class.
Just a quick note on this one, Markus (I think) added the webgraph commands to the nutch script so this creates a simpler working environment from 1.4 onwards. On Thu, Sep 22, 2011 at 7:53 AM, Thomas Anderson <[email protected]>wrote: > I follow the example tutorial at > http://wiki.apache.org/nutch/NewScoringIndexingExample. Nearly all > command executes well except LinkRank command. > > When executing LinkRank command `nutch > org.apache.nutch.scoring.webgraph.LinkRank -webgraphdb > crawl/webgraphdb/,` it throws following exception. > > 11/09/22 14:44:56 FATAL webgraph.LinkRank: LinkAnalysis: > java.io.IOException: No links to process, is the webgraph empty? > at > org.apache.nutch.scoring.webgraph.LinkRank.runCounter(LinkRank.java:131) > at > org.apache.nutch.scoring.webgraph.LinkRank.analyze(LinkRank.java:610) > at org.apache.nutch.scoring.webgraph.LinkRank.run(LinkRank.java:686) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > at > org.apache.nutch.scoring.webgraph.LinkRank.main(LinkRank.java:656) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > at java.lang.reflect.Method.invoke(Method.java:597) > at org.apache.hadoop.util.RunJar.main(RunJar.java:156) > > At beginning I do not add the following properties to > hadoop/conf/nutch-site.xml > > <!-- linkrank scoring properties --> > <property> > <name>link.ignore.internal.host</name> > <value>true</value> > <description>Ignore outlinks to the same hostname.</description> > </property> > > <property> > <name>link.ignore.internal.domain</name> > <value>true</value> > <description>Ignore outlinks to the same domain.</description> > </property> > > <property> > <name>link.ignore.limit.page</name> > <value>true</value> > <description>Limit to only a single outlink to the same > page.</description> > </property> > > <property> > <name>link.ignore.limit.domain</name> > <value>true</value> > <description>Limit to only a single outlink to the same > domain.</description> > </property> > > But after adding those properties, the exception remains. > What may cause such error? > > Environment: java "1.6.0_26", debian with 2.6.39-2-686-pae kernel, > nutch 1.3, hadoop 0.20.2 > > Thanks > -- *Lewis*

