The same now occasionally happens with the crawldb. Today we tested new 
filters a couple of times. The job failed without failure twice, no strange 
error message or whatever. The only clue is that the tmp files does not 
replace the `current` dir.

Anything?


> Hi,
> 
> Last night i started another round of link inverting and merging with the
> existing LinkDB. Many hours later i see the job finished without failures
> or errors.
> 
> Yet the LinkDB wasn't updated. To my surprise i saw the intermediate files
> hanging around in my Hadoop ~ but i expected the temp files to be cleaned
> and the new LinkDB file to be copied over the old one.
> 
> The new file that ought to be copied over is also broken. I cannot
> readlinkdb. Gladly the old still works but before i try to debug some
> process that takes many hours.. any idea?
> 
> Thanks

Reply via email to