since it is complaining about logging, check if the path log4j is trying
to access is valid and the hadoop user has permissions to access it.

>
> i can't solve it now
>
> jibjoice wrote:
>>
>> i follow this link "http://wiki.apache.org/nutch/NutchHadoopTutorial"; so
>> i
>> think it's not about the conf/crawl-urlfilter.txt file when i use this
>> command "bin/nutch crawl urls -dir crawled -depth 3" again it shows :
>> Generator: Selecting best-scoring urls due for fetch.
>> Generator: starting
>> Generator: segment: crawled/segments/25501221110712
>> Generator: filtering: false
>> Generator: topN: 2147483647
>> Generator: Partitioning selected urls by host, for politeness.
>> Generator: done.
>> Fetcher: starting
>> Fetcher: segment: crawled/segments/25501221110712
>> Fetcher: done
>> CrawlDb update: starting
>> CrawlDb update: db: crawled/crawldb
>> CrawlDb update: segments: [crawled/segments/25501221110712]
>> CrawlDb update: additions allowed: true
>> CrawlDb update: URL normalizing: true
>> CrawlDb update: URL filtering: true
>> CrawlDb update: Merging segment data into db.
>> CrawlDb update: done
>> Generator: Selecting best-scoring urls due for fetch.
>> Generator: starting
>> Generator: segment: crawled/segments/25501221110908
>> Generator: filtering: false
>> Generator: topN: 2147483647
>> Generator: Partitioning selected urls by host, for politeness.
>> Generator: done.
>> Fetcher: starting
>> Fetcher: segment: crawled/segments/25501221110908
>> Fetcher: done
>> CrawlDb update: starting
>> CrawlDb update: db: crawled/crawldb
>> CrawlDb update: segments: [crawled/segments/25501221110908]
>> CrawlDb update: additions allowed: true
>> CrawlDb update: URL normalizing: true
>> CrawlDb update: URL filtering: true
>> CrawlDb update: Merging segment data into db.
>> CrawlDb update: done
>> LinkDb: starting
>> LinkDb: linkdb: crawled/linkdb
>> LinkDb: URL normalize: true
>> LinkDb: URL filter: true
>> LinkDb: adding segment: /user/nutch/crawled/segments/25501221110519
>> LinkDb: adding segment: /user/nutch/crawled/segments/25501221110712
>> LinkDb: adding segment: /user/nutch/crawled/segments/25501221110908
>> LinkDb: done
>> Indexer: starting
>> Indexer: linkdb: crawled/linkdb
>> Indexer: adding segment: /user/nutch/crawled/segments/25501221110519
>> Indexer: adding segment: /user/nutch/crawled/segments/25501221110712
>> Indexer: adding segment: /user/nutch/crawled/segments/25501221110908
>> Indexer: done
>> Dedup: starting
>> Dedup: adding indexes in: crawled/indexes
>> task_0017_m_000000_0: log4j:ERROR Either File or DatePattern options are
>> not set for appender [DRFA].
>> task_0017_m_000001_0: log4j:ERROR setFile(null,true) call failed.
>> task_0017_m_000001_0: java.io.FileNotFoundException: /nutch/search/logs
>> (Is a directory)
>> task_0017_m_000001_0:   at java.io.FileOutputStream.openAppend(Native
>> Method)
>> task_0017_m_000001_0:   at
>> java.io.FileOutputStream.<init>(FileOutputStream.java:177)
>> task_0017_m_000001_0:   at
>> java.io.FileOutputStream.<init>(FileOutputStream.java:102)
>> task_0017_m_000001_0:   at
>> org.apache.log4j.FileAppender.setFile(FileAppender.java:289)
>> task_0017_m_000001_0:   at
>> org.apache.log4j.FileAppender.activateOptions(FileAppender.java:163)
>> task_0017_m_000001_0:   at
>> org.apache.log4j.DailyRollingFileAppender.activateOptions(DailyRollingFileAppender.java:215)
>> task_0017_m_000001_0:   at
>> org.apache.log4j.config.PropertySetter.activate(PropertySetter.java:256)
>> task_0017_m_000001_0:   at
>> org.apache.log4j.config.PropertySetter.setProperties(PropertySetter.java:132)
>> task_0017_m_000001_0:   at
>> org.apache.log4j.config.PropertySetter.setProperties(PropertySetter.java:96)
>> task_0017_m_000001_0:   at
>> org.apache.log4j.PropertyConfigurator.parseAppender(PropertyConfigurator.java:654)
>> task_0017_m_000001_0:   at
>> org.apache.log4j.PropertyConfigurator.parseCategory(PropertyConfigurator.java:612)
>> task_0017_m_000001_0:   at
>> org.apache.log4j.PropertyConfigurator.configureRootCategory(PropertyConfigurator.java:509)
>> task_0017_m_000001_0:   at
>> org.apache.log4j.PropertyConfigurator.doConfigure(PropertyConfigurator.java:415)
>> task_0017_m_000001_0:   at
>> org.apache.log4j.PropertyConfigurator.doConfigure(PropertyConfigurator.java:441)
>> task_0017_m_000001_0:   at
>> org.apache.log4j.helpers.OptionConverter.selectAndConfigure(OptionConverter.java:468)
>> task_0017_m_000001_0:   at
>> org.apache.log4j.LogManager.<clinit>(LogManager.java:122)
>> task_0017_m_000001_0:   at
>> org.apache.log4j.Logger.getLogger(Logger.java:104)
>> task_0017_m_000001_0:   at
>> org.apache.commons.logging.impl.Log4JLogger.getLogger(Log4JLogger.java:229)
>> task_0017_m_000001_0:   at
>> org.apache.commons.logging.impl.Log4JLogger.<init>(Log4JLogger.java:65)
>> task_0017_m_000001_0:   at
>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>> task_0017_m_000001_0:   at
>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
>> task_0017_m_000001_0:   at
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>> task_0017_m_000001_0:   at
>> java.lang.reflect.Constructor.newInstance(Constructor.java:513)
>> task_0017_m_000001_0:   at
>> org.apache.commons.logging.impl.LogFactoryImpl.newInstance(LogFactoryImpl.java:529)
>> task_0017_m_000001_0:   at
>> org.apache.commons.logging.impl.LogFactoryImpl.getInstance(LogFactoryImpl.java:235)
>> task_0017_m_000001_0:   at
>> org.apache.commons.logging.LogFactory.getLog(LogFactory.java:370)
>> task_0017_m_000001_0:   at
>> org.apache.hadoop.mapred.TaskTracker.<clinit>(TaskTracker.java:82)
>> task_0017_m_000001_0:   at
>> org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:1423)
>> task_0017_m_000001_0: log4j:ERROR Either File or DatePattern options are
>> not set for appender [DRFA].
>> task_0017_m_000000_1: log4j:ERROR setFile(null,true) call failed.
>> task_0017_m_000000_1: java.io.FileNotFoundException: /nutch/search/logs
>> (Is a directory)
>> task_0017_m_000000_1:   at java.io.FileOutputStream.openAppend(Native
>> Method)
>> task_0017_m_000000_1:   at
>> java.io.FileOutputStream.<init>(FileOutputStream.java:177)
>> task_0017_m_000000_1:   at
>> java.io.FileOutputStream.<init>(FileOutputStream.java:102)
>> task_0017_m_000000_1:   at
>> org.apache.log4j.FileAppender.setFile(FileAppender.java:289)
>> task_0017_m_000000_1:   at
>> org.apache.log4j.FileAppender.activateOptions(FileAppender.java:163)
>> task_0017_m_000000_1:   at
>> org.apache.log4j.DailyRollingFileAppender.activateOptions(DailyRollingFileAppender.java:215)
>> task_0017_m_000000_1:   at
>> org.apache.log4j.config.PropertySetter.activate(PropertySetter.java:256)
>> task_0017_m_000000_1:   at
>> org.apache.log4j.config.PropertySetter.setProperties(PropertySetter.java:132)
>> task_0017_m_000000_1:   at
>> org.apache.log4j.config.PropertySetter.setProperties(PropertySetter.java:96)
>> task_0017_m_000000_1:   at
>> org.apache.log4j.PropertyConfigurator.parseAppender(PropertyConfigurator.java:654)
>> task_0017_m_000000_1:   at
>> org.apache.log4j.PropertyConfigurator.parseCategory(PropertyConfigurator.java:612)
>> task_0017_m_000000_1:   at
>> org.apache.log4j.PropertyConfigurator.configureRootCategory(PropertyConfigurator.java:509)
>> task_0017_m_000000_1:   at
>> org.apache.log4j.PropertyConfigurator.doConfigure(PropertyConfigurator.java:415)
>> task_0017_m_000000_1:   at
>> org.apache.log4j.PropertyConfigurator.doConfigure(PropertyConfigurator.java:441)
>> task_0017_m_000000_1:   at
>> org.apache.log4j.helpers.OptionConverter.selectAndConfigure(OptionConverter.java:468)
>> task_0017_m_000000_1:   at
>> org.apache.log4j.LogManager.<clinit>(LogManager.java:122)
>> task_0017_m_000000_1:   at
>> org.apache.log4j.Logger.getLogger(Logger.java:104)
>> task_0017_m_000000_1:   at
>> org.apache.commons.logging.impl.Log4JLogger.getLogger(Log4JLogger.java:229)
>> task_0017_m_000000_1:   at
>> org.apache.commons.logging.impl.Log4JLogger.<init>(Log4JLogger.java:65)
>> task_0017_m_000000_1:   at
>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>> task_0017_m_000000_1:   at
>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
>> task_0017_m_000000_1:   at
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>> task_0017_m_000000_1:   at
>> java.lang.reflect.Constructor.newInstance(Constructor.java:513)
>> task_0017_m_000000_1:   at
>> org.apache.commons.logging.impl.LogFactoryImpl.newInstance(LogFactoryImpl.java:529)
>> task_0017_m_000000_1:   at
>> org.apache.commons.logging.impl.LogFactoryImpl.getInstance(LogFactoryImpl.java:235)
>> task_0017_m_000000_3:   at
>> org.apache.commons.logging.LogFactory.getLog(LogFactory.java:370)
>> task_0017_m_000000_3:   at
>> org.apache.hadoop.mapred.TaskTracker.<clinit>(TaskTracker.java:82)
>> task_0017_m_000000_3:   at
>> org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:1423)
>> task_0017_m_000000_3: log4j:ERROR Either File or DatePattern options are
>> not set for appender [DRFA].
>> Exception in thread "main" java.io.IOException: Job failed!
>>         at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:604)
>>         at
>> org.apache.nutch.indexer.DeleteDuplicates.dedup(DeleteDuplicates.java:439)
>>         at org.apache.nutch.crawl.Crawl.main(Crawl.java:135)
>>
>> i don't know what happen?
>>
>>
>> pvvpr wrote:
>>>
>>> I think you need to check the conf/crawl-urlfilter.txt file
>>>
>>> On Thursday 20 December 2007 04:55, jibjoice wrote:
>>>> please, help me to solve it
>>>>
>>>> jibjoice wrote:
>>>> > where i should solve this? why it generated 0 records?
>>>> >
>>>> > pvvpr wrote:
>>>> >> basically your indexes are empty since no URLs were generated and
>>>> >> fetched. See
>>>> >> this,
>>>> >>
>>>> >>> > - Generator: 0 records selected for fetching, exiting ...
>>>> >>> > - Stopping at depth=0 - no more URLs to fetch.
>>>> >>> > - No URLs to fetch - check your seed list and URL filters.
>>>> >>> > - crawl finished: crawled
>>>> >>
>>>> >> when no pages are indexed, dedup throws Exception
>>>> >>
>>>> >> On Tuesday 18 December 2007 21:33, jibjoice wrote:
>>>> >>> i can't solve it now, pls help me
>>>> >>>
>>>> >>> jibjoice wrote:
>>>> >>> > i use nutch-0.9, hadoop-0.12.2 and i use this command "bin/nutch
>>>> >>> > crawl urls -dir crawled -depth 3" have error :
>>>> >>> >
>>>> >>> > - crawl started in: crawled
>>>> >>> > - rootUrlDir = input
>>>> >>> > - threads = 10
>>>> >>> > - depth = 3
>>>> >>> > - Injector: starting
>>>> >>> > - Injector: crawlDb: crawled/crawldb
>>>> >>> > - Injector: urlDir: input
>>>> >>> > - Injector: Converting injected urls to crawl db entries.
>>>> >>> > - Total input paths to process : 1
>>>> >>> > - Running job: job_0001
>>>> >>> > - map 0% reduce 0%
>>>> >>> > - map 100% reduce 0%
>>>> >>> > - map 100% reduce 100%
>>>> >>> > - Job complete: job_0001
>>>> >>> > - Counters: 6
>>>> >>> > - Map-Reduce Framework
>>>> >>> > - Map input records=3
>>>> >>> > - Map output records=1
>>>> >>> > - Map input bytes=22
>>>> >>> > - Map output bytes=52
>>>> >>> > - Reduce input records=1
>>>> >>> > - Reduce output records=1
>>>> >>> > - Injector: Merging injected urls into crawl db.
>>>> >>> > - Total input paths to process : 2
>>>> >>> > - Running job: job_0002
>>>> >>> > - map 0% reduce 0%
>>>> >>> > - map 100% reduce 0%
>>>> >>> > - map 100% reduce 58%
>>>> >>> > - map 100% reduce 100%
>>>> >>> > - Job complete: job_0002
>>>> >>> > - Counters: 6
>>>> >>> > - Map-Reduce Framework
>>>> >>> > - Map input records=3
>>>> >>> > - Map output records=1
>>>> >>> > - Map input bytes=60
>>>> >>> > - Map output bytes=52
>>>> >>> > - Reduce input records=1
>>>> >>> > - Reduce output records=1
>>>> >>> > - Injector: done
>>>> >>> > - Generator: Selecting best-scoring urls due for fetch.
>>>> >>> > - Generator: starting
>>>> >>> > - Generator: segment: crawled/segments/25501213164325
>>>> >>> > - Generator: filtering: false
>>>> >>> > - Generator: topN: 2147483647
>>>> >>> > - Total input paths to process : 2
>>>> >>> > - Running job: job_0003
>>>> >>> > - map 0% reduce 0%
>>>> >>> > - map 100% reduce 0%
>>>> >>> > - map 100% reduce 100%
>>>> >>> > - Job complete: job_0003
>>>> >>> > - Counters: 6
>>>> >>> > - Map-Reduce Framework
>>>> >>> > - Map input records=3
>>>> >>> > - Map output records=1
>>>> >>> > - Map input bytes=59
>>>> >>> > - Map output bytes=77
>>>> >>> > - Reduce input records=1
>>>> >>> > - Reduce output records=1
>>>> >>> > - Generator: 0 records selected for fetching, exiting ...
>>>> >>> > - Stopping at depth=0 - no more URLs to fetch.
>>>> >>> > - No URLs to fetch - check your seed list and URL filters.
>>>> >>> > - crawl finished: crawled
>>>> >>> >
>>>> >>> > but sometime i crawl some url it has error indexes time that
>>>> >>> >
>>>> >>> > - Indexer: done
>>>> >>> > - Dedup: starting
>>>> >>> > - Dedup: adding indexes in: crawled/indexes
>>>> >>> > - Total input paths to process : 2
>>>> >>> > - Running job: job_0025
>>>> >>> > - map 0% reduce 0%
>>>> >>> > - Task Id : task_0025_m_000001_0, Status : FAILED
>>>> >>> > task_0025_m_000001_0: - Error running child
>>>> >>> > task_0025_m_000001_0: java.lang.ArrayIndexOutOfBoundsException:
>>>> -1
>>>> >>> > task_0025_m_000001_0: at
>>>> >>> >
>>>> org.apache.lucene.index.MultiReader.isDeleted(MultiReader.java:113)
>>>> >>> > task_0025_m_000001_0: at
>>>> >>> >
>>>> org.apache.nutch.indexer.DeleteDuplicates$InputFormat$DDRecordReade
>>>> >>> > r.next(DeleteDuplicates.java:176)
>>>> >>> > task_0025_m_000001_0: at
>>>> >>> > org.apache.hadoop.mapred.MapTask$1.next(MapTask.java:157)
>>>> >>> > task_0025_m_000001_0: at
>>>> >>> > org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:46)
>>>> >>> > task_0025_m_000001_0: at org.apache.hadoop.mapred.MapTask.run
>>>> >>> > (MapTask.java:175)
>>>> >>> > task_0025_m_000001_0: at
>>>> >>> > org.apache.hadoop.mapred.TaskTracker$Child.main
>>>> >>> > (TaskTracker.java:1445)
>>>> >>> > - Task Id : task_0025_m_000000_0, Status : FAILED
>>>> >>> > task_0025_m_000000_0: - Error running child
>>>> >>> > task_0025_m_000000_0: java.lang.ArrayIndexOutOfBoundsException:
>>>> -1
>>>> >>> > task_0025_m_000000_0: at
>>>> >>> >
>>>> org.apache.lucene.index.MultiReader.isDeleted(MultiReader.java:113)
>>>> >>> > task_0025_m_000000_0: at
>>>> >>> >
>>>> org.apache.nutch.indexer.DeleteDuplicates$InputFormat$DDRecordReade
>>>> >>> > r.next(DeleteDuplicates.java:176)
>>>> >>> > task_0025_m_000000_0: at
>>>> >>> > org.apache.hadoop.mapred.MapTask$1.next(MapTask.java:157)
>>>> >>> > task_0025_m_000000_0: at
>>>> >>> > org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:46)
>>>> >>> > task_0025_m_000000_0: at org.apache.hadoop.mapred.MapTask.run
>>>> >>> > (MapTask.java:175)
>>>> >>> > task_0025_m_000000_0: at
>>>> >>> > org.apache.hadoop.mapred.TaskTracker$Child.main
>>>> >>> > (TaskTracker.java:1445)
>>>> >>> > - Task Id : task_0025_m_000000_1, Status : FAILED
>>>> >>> > task_0025_m_000000_1: - Error running child
>>>> >>> > task_0025_m_000000_1: java.lang.ArrayIndexOutOfBoundsException:
>>>> -1
>>>> >>> > task_0025_m_000000_1: at
>>>> >>> >
>>>> org.apache.lucene.index.MultiReader.isDeleted(MultiReader.java:113)
>>>> >>> > task_0025_m_000000_1: at
>>>> >>> >
>>>> org.apache.nutch.indexer.DeleteDuplicates$InputFormat$DDRecordReade
>>>> >>> > r.next(DeleteDuplicates.java:176)
>>>> >>> > task_0025_m_000000_1: at
>>>> >>> > org.apache.hadoop.mapred.MapTask$1.next(MapTask.java:157)
>>>> >>> > task_0025_m_000000_1: at
>>>> >>> > org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:46)
>>>> >>> > task_0025_m_000000_1: at org.apache.hadoop.mapred.MapTask.run
>>>> >>> > (MapTask.java:175)
>>>> >>> > task_0025_m_000000_1: at
>>>> >>> > org.apache.hadoop.mapred.TaskTracker$Child.main
>>>> >>> > (TaskTracker.java:1445)
>>>> >>> > - Task Id : task_0025_m_000001_1, Status : FAILED
>>>> >>> > task_0025_m_000001_1: - Error running child
>>>> >>> > task_0025_m_000001_1: java.lang.ArrayIndexOutOfBoundsException:
>>>> -1
>>>> >>> > task_0025_m_000001_1: at
>>>> >>> >
>>>> org.apache.lucene.index.MultiReader.isDeleted(MultiReader.java:113)
>>>> >>> > task_0025_m_000001_1: at
>>>> >>> >
>>>> org.apache.nutch.indexer.DeleteDuplicates$InputFormat$DDRecordReade
>>>> >>> > r.next(DeleteDuplicates.java:176)
>>>> >>> > task_0025_m_000001_1: at
>>>> >>> > org.apache.hadoop.mapred.MapTask$1.next(MapTask.java:157)
>>>> >>> > task_0025_m_000001_1: at
>>>> >>> > org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:46)
>>>> >>> > task_0025_m_000001_1: at org.apache.hadoop.mapred.MapTask.run
>>>> >>> > (MapTask.java:175)
>>>> >>> > task_0025_m_000001_1: at
>>>> >>> > org.apache.hadoop.mapred.TaskTracker$Child.main
>>>> >>> > (TaskTracker.java:1445)
>>>> >>> > - Task Id : task_0025_m_000001_2, Status : FAILED
>>>> >>> > task_0025_m_000001_2: - Error running child
>>>> >>> > task_0025_m_000001_2: java.lang.ArrayIndexOutOfBoundsException:
>>>> -1
>>>> >>> > task_0025_m_000001_2: at
>>>> >>> >
>>>> org.apache.lucene.index.MultiReader.isDeleted(MultiReader.java:113)
>>>> >>> > task_0025_m_000001_2: at
>>>> >>> >
>>>> org.apache.nutch.indexer.DeleteDuplicates$InputFormat$DDRecordReade
>>>> >>> > r.next(DeleteDuplicates.java:176)
>>>> >>> > task_0025_m_000001_2: at
>>>> >>> > org.apache.hadoop.mapred.MapTask$1.next(MapTask.java:157)
>>>> >>> > task_0025_m_000001_2: at
>>>> >>> > org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:46)
>>>> >>> > task_0025_m_000001_2: at org.apache.hadoop.mapred.MapTask.run
>>>> >>> > (MapTask.java:175)
>>>> >>> > task_0025_m_000001_2: at
>>>> >>> > org.apache.hadoop.mapred.TaskTracker$Child.main
>>>> >>> > (TaskTracker.java:1445)
>>>> >>> > - Task Id : task_0025_m_000000_2, Status : FAILED
>>>> >>> > task_0025_m_000000_2: - Error running child
>>>> >>> > task_0025_m_000000_2: java.lang.ArrayIndexOutOfBoundsException:
>>>> -1
>>>> >>> > task_0025_m_000000_2: at
>>>> >>> >
>>>> org.apache.lucene.index.MultiReader.isDeleted(MultiReader.java:113)
>>>> >>> > task_0025_m_000000_2: at
>>>> >>> >
>>>> org.apache.nutch.indexer.DeleteDuplicates$InputFormat$DDRecordReade
>>>> >>> > r.next(DeleteDuplicates.java:176)
>>>> >>> > task_0025_m_000000_2: at
>>>> >>> > org.apache.hadoop.mapred.MapTask$1.next(MapTask.java:157)
>>>> >>> > task_0025_m_000000_2: at
>>>> >>> > org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:46)
>>>> >>> > task_0025_m_000000_2: at org.apache.hadoop.mapred.MapTask.run
>>>> >>> > (MapTask.java:175)
>>>> >>> > task_0025_m_000000_2: at
>>>> >>> > org.apache.hadoop.mapred.TaskTracker$Child.main
>>>> >>> > (TaskTracker.java:1445)
>>>> >>> > - map 100% reduce 100%
>>>> >>> > - Task Id : task_0025_m_000001_3, Status : FAILED
>>>> >>> > task_0025_m_000001_3: - Error running child
>>>> >>> > task_0025_m_000001_3: java.lang.ArrayIndexOutOfBoundsException:
>>>> -1
>>>> >>> > task_0025_m_000001_3: at
>>>> >>> >
>>>> org.apache.lucene.index.MultiReader.isDeleted(MultiReader.java:113)
>>>> >>> > task_0025_m_000001_3: at
>>>> >>> >
>>>> org.apache.nutch.indexer.DeleteDuplicates$InputFormat$DDRecordReade
>>>> >>> > r.next(DeleteDuplicates.java:176)
>>>> >>> > task_0025_m_000001_3: at
>>>> >>> > org.apache.hadoop.mapred.MapTask$1.next(MapTask.java:157)
>>>> >>> > task_0025_m_000001_3: at
>>>> >>> > org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:46)
>>>> >>> > task_0025_m_000001_3: at org.apache.hadoop.mapred.MapTask.run
>>>> >>> > (MapTask.java:175)
>>>> >>> > task_0025_m_000001_3: at
>>>> >>> > org.apache.hadoop.mapred.TaskTracker$Child.main
>>>> >>> > (TaskTracker.java:1445)
>>>> >>> > - Task Id : task_0025_m_000000_3, Status : FAILED
>>>> >>> > task_0025_m_000000_3: - Error running child
>>>> >>> > task_0025_m_000000_3: java.lang.ArrayIndexOutOfBoundsException:
>>>> -1
>>>> >>> > task_0025_m_000000_3: at
>>>> >>> >
>>>> org.apache.lucene.index.MultiReader.isDeleted(MultiReader.java:113)
>>>> >>> > task_0025_m_000000_3: at
>>>> >>> >
>>>> org.apache.nutch.indexer.DeleteDuplicates$InputFormat$DDRecordReade
>>>> >>> > r.next(DeleteDuplicates.java:176)
>>>> >>> > task_0025_m_000000_3: at
>>>> >>> > org.apache.hadoop.mapred.MapTask$1.next(MapTask.java:157)
>>>> >>> > task_0025_m_000000_3: at
>>>> >>> > org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:46)
>>>> >>> > task_0025_m_000000_3: at org.apache.hadoop.mapred.MapTask.run
>>>> >>> > (MapTask.java:175)
>>>> >>> > task_0025_m_000000_3: at
>>>> >>> > org.apache.hadoop.mapred.TaskTracker$Child.main
>>>> >>> > (TaskTracker.java:1445)
>>>> >>> > Exception in thread "main" java.io.IOException: Job failed!
>>>> >>> > at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:604)
>>>> >>> > at org.apache.nutch.indexer.DeleteDuplicates.dedup
>>>> >>> > (DeleteDuplicates.java:439)
>>>> >>> > at org.apache.nutch.crawl.Crawl.main(Crawl.java:135)
>>>> >>> >
>>>> >>> > how i solve it?
>>>
>>>
>>>
>>>
>>
>>
>
> --
> View this message in context:
> http://www.nabble.com/Nutch-crawl-problem-tp14327978p14492766.html
> Sent from the Hadoop Users mailing list archive at Nabble.com.
>
>
>



Reply via email to