Hi alexei.
Make sure about markus suggestion, i had a same problem with /tmp folder space 
while nutch is crawling. This folder is cleaned when you reboot the system, but 
nutch check the available space and it can throw exceptions. verify the space 
with 
du -hs /tmp/
also check this thread 
http://mail-archives.apache.org/mod_mbox/nutch-user/201302.mbox/browser





----- Mensaje original -----
De: "Alexei Korolev" <[email protected]>
Para: [email protected]
Enviados: Lunes, 11 de Febrero 2013 3:40:06
Asunto: Re: DiskChecker$DiskErrorException

Hi,

Yes

Filesystem           1K-blocks      Used Available Use% Mounted on
/dev/md2             1065281580 592273404 419321144  59% /
udev                   8177228         8   8177220   1% /dev
tmpfs                  3274592       328   3274264   1% /run
none                      5120         0      5120   0% /run/lock
none                   8186476         0   8186476   0% /run/shm
/dev/md3             1808084492  15283960 1701678392   1% /home
/dev/md1                507684     38099    443374   8% /boot

On Mon, Feb 11, 2013 at 12:33 PM, Markus Jelsma
<[email protected]>wrote:

> Hi- Also enough space in your /tmp directory?
>
> Cheers
>
>
>
> -----Original message-----
> > From:Alexei Korolev <[email protected]>
> > Sent: Mon 11-Feb-2013 09:27
> > To: [email protected]
> > Subject: DiskChecker$DiskErrorException
> >
> > Hello,
> >
> > Already twice I got this error:
> >
> > 2013-02-08 15:26:11,674 WARN  mapred.LocalJobRunner - job_local_0001
> > org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find
> >
> taskTracker/jobcache/job_local_0001/attempt_local_0001_m_000000_0/output/spill0.out
> > in any of the configur
> > ed local directories
> >         at
> >
> org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathToRead(LocalDirAllocator.java:389)
> >         at
> >
> org.apache.hadoop.fs.LocalDirAllocator.getLocalPathToRead(LocalDirAllocator.java:138)
> >         at
> >
> org.apache.hadoop.mapred.MapOutputFile.getSpillFile(MapOutputFile.java:94)
> >         at
> >
> org.apache.hadoop.mapred.MapTask$MapOutputBuffer.mergeParts(MapTask.java:1443)
> >         at
> > org.apache.hadoop.mapred.MapTask$MapOutputBuffer.flush(MapTask.java:1154)
> >         at
> org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:359)
> >         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:307)
> >         at
> > org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:177)
> > 2013-02-08 15:26:12,515 ERROR fetcher.Fetcher - Fetcher:
> > java.io.IOException: Job failed!
> >         at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1252)
> >         at org.apache.nutch.fetcher.Fetcher.fetch(Fetcher.java:1204)
> >         at org.apache.nutch.fetcher.Fetcher.run(Fetcher.java:1240)
> >         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
> >         at org.apache.nutch.fetcher.Fetcher.main(Fetcher.java:1213)
> >
> > I've checked in google, but no luck. I run nutch 1.4 locally and have a
> > plenty of free space on disk.
> > I would much appreciate for some help.
> >
> > Thanks.
> >
> >
> > --
> > Alexei A. Korolev
> >
>



-- 
Alexei A. Korolev

Reply via email to