http://   ?

hth

On Fri, Aug 3, 2012 at 9:53 AM, Alexei Korolev <alexei.koro...@gmail.com> wrote:
> Hello,
>
> I have small script
>
> $NUTCH_PATH inject crawl/crawldb seed.txt
> $NUTCH_PATH generate crawl/crawldb crawl/crawldb/segments -adddays 0
>
> s1=`ls -d crawl/crawldb/segments/* | tail -1`
> $NUTCH_PATH fetch $s1
> $NUTCH_PATH parse $s1
> $NUTCH_PATH updatedb crawl/crawldb $s1
>
> In seed.txt I have just one site, for example "test.com". When I start
> script it falls on fetch phase.
> If I change test.com on www.test.com it works fine. Seems the reason, that
> outgoing link on test.com all have www. prefix.
> What I need to change in nutch config for work with test.com?
>
> Thank you in advance. I hope my explanation is clear :)
>
> --
> Alexei A. Korolev



-- 
Lewis

Reply via email to