Hi, have you solved the problem? I have the same issue and I can not read the link db which I think is wrong because the generated file (linkdb directory) are all 1kb. Using the 1.0 version of Nutch
Thanks 2010/4/14 Norman Birke <[email protected]> > > Hi, > > I am trying to dump my linkdb content for analysis using the following > command: > bin/nutch readlinkdb crawl/linkdb -dump readlinkdb_dump > > I receive the following output in my shell: > LinkDb dump: starting > LinkDb db: crawl/linkdb/ > > After that the readlinkdb_dump folder exists and in it the 2 files: > .part-00000.crc (which has a size of 8 byte) > part-00000 (which has a size of 0 byte) > > As I have 686 URLs in my crawldb the file size seems a bit small to me. > Everything else works fine - I can read and dump my crawldb, read and dump > segments and so on. > > Any idea what might be messed up? > > Many thanks in advance, > Norman >

