Hi,
We've been experiencing unpredictable nightly failures for both Nutch trunk
over the last while. One of our memory intensive tests pops up as the
common denominator with the following output.
2011-11-03 04:17:37,478 WARN mapred.LocalJobRunner
(LocalJobRunner.java:run(256)) - job_local_0001
org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find
taskTracker/jobcache/job_local_0001/attempt_local_0001_m_000000_0/output/file.out
in any of the configured local directories
at
org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathToRead(LocalDirAllocator.java:389)
at
org.apache.hadoop.fs.LocalDirAllocator.getLocalPathToRead(LocalDirAllocator.java:138)
at
org.apache.hadoop.mapred.MapOutputFile.getOutputFile(MapOutputFile.java:50)
at
org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:193)
Between us we've agreed that the DiskCheckerException usually smells like
running out of disk space in the designated tmp dir, therefore it has been
suggested that we configure the build to build on any given machine from a
number of available slaves. Can anyone please provide a bit of insight as
to whether this is a suitable thing to do, and secondly how I can go about
configuring the builds as per my comments.
Thank you
--
*Lewis*