Hello everybody. I have one problem with my current Nutch 2.1 configuration. I run my crawling process (throught bin/crawl script) every morning during the last few months and everything went fine - I mean that in every run the script generated new batch which was fetched and after that parsed and finally indexed into Solr (I use also hBase as my data store).
I do not know why but something had broken and nothing is fetched now (also because this problem there is nothing to parse and index). If it could be helpful, I tried to do bin/nutch readdb -stats and output is here: http://pastebin.com/r3vFK4T4 <http://pastebin.com/r3vFK4T4> There are some things like status 3 (status_gone): 9907 status 1 (status_unfetched): 6798166 status 0 (null): 2 but I dont understand too much this attributes... but 6798166 unfetched urls seem very suspicious. During crawling process I can see that GeneratorJob generates some batch id but I dont know if this batchId is not empty or something. /GeneratorJob: generated batch id: 1329930901-1268252438 FetcherJob: starting FetcherJob : timelimit set for : -1 FetcherJob: threads: 10 FetcherJob: parsing: false FetcherJob: resuming: false FetcherJob: batchId: 1329930901-1268252438 Using queue mode : byHost . . . -activeThreads=0, spinWaiting=0, fetchQueues= 0, fetchQueues.totalSize=0 -activeThreads=0 FetcherJob: done / fetching is really fast because as I said there is no url to be fetched. Can you help me or give any advice? Thanks, Jan -- View this message in context: http://lucene.472066.n3.nabble.com/Nutch-2-1-fetching-is-not-working-maybe-broken-generate-tp4123813.html Sent from the Nutch - User mailing list archive at Nabble.com.

