Tejas Patil wrote
> On Thu, Mar 14, 2013 at 7:36 PM, Dat Tran <

> tranquydat@

> > wrote:
> 
>> Update: When i remove or add new url in urls.txt( seed list), it is
>> strange
>> the crawling result is not changed. It means  nutch crawls always the
>> first
>> seed list.
> 
> 
> You mean the older copy of seeds file is used effectively and the updates
> you did not got reflected ?
> What command are you using to run the crawler ?
> Yes i used the older copy of seeds file, updated and did not got
> reflected. Even i uses the other seed file, it is not reflected.
> To run the crawler i execute the command :  bin/nutch parse  urls ( urls
> is the directory where seed file located)
> This problem is caused by any temporary file ? How can i resolve
>> it? where can i find temporary file of nutch
>>
> 
> In shell, you can use "ls -a" to see the hidden files. That way you can
> find if there are hidden backup files in your seeds directory.
> 
>>
>>
>> --
>> View this message in context:
>> http://lucene.472066.n3.nabble.com/Iterative-Crawling-tp4046501p4047572.html
>> Sent from the Nutch - User mailing list archive at Nabble.com.
>>





--
View this message in context: 
http://lucene.472066.n3.nabble.com/Iterative-Crawling-tp4046501p4047643.html
Sent from the Nutch - User mailing list archive at Nabble.com.

Reply via email to