After crawling when I run that command:

bin/nutch solrindex http://localhost:8983/solr -index

Sometims I get that error:

FetcherJob: threads: 10                                                         
                                                                                
                             
FetcherJob: parsing: false                                                      
                                                                                
                             
FetcherJob: resuming: false                                                     
                                                                                
                             
FetcherJob : timelimit set for : -1                                             
                                                                                
                             
Exception in thread "main" java.lang.RuntimeException: job failed:
name=fetch, jobid=job_local_0031
        at
org.apache.nutch.util.NutchJob.waitForCompletion(NutchJob.java:54)
        at org.apache.nutch.fetcher.FetcherJob.run(FetcherJob.java:194)
        at org.apache.nutch.crawl.Crawler.runTool(Crawler.java:68)
        at org.apache.nutch.crawl.Crawler.run(Crawler.java:161)
        at org.apache.nutch.crawl.Crawler.run(Crawler.java:250)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
        at org.apache.nutch.crawl.Crawler.main(Crawler.java:257)

When I check hadoop log I see that:

2013-03-24 01:58:42,151 WARN  mapred.LocalJobRunner - job_local_0031            
                                                                                
                             
java.lang.OutOfMemoryError: unable to create new native thread                  
                                                                                
                             
        at java.lang.Thread.start0(Native Method)                               
                                                                                
                             
        at java.lang.Thread.start(Thread.java:691)                              
                                                                                
                             
        at
org.apache.nutch.fetcher.FetcherReducer.run(FetcherReducer.java:790)            
                                                                                
                  
        at
org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:649)          
                                                                                
                  
        at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:417)         
                                                                                
                             
        at
org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:260) 

I have Nutch 2.1, Solr 4.2 running on a core i7 64 bit centos that has 8
core (4 real core) and 8 GB Ram, at CentOS 6.4 What may be the reason?



--
View this message in context: 
http://lucene.472066.n3.nabble.com/waitForCompletion-Error-tp4050809.html
Sent from the Nutch - User mailing list archive at Nabble.com.

Reply via email to