[ 
https://issues.apache.org/jira/browse/HADOOP-573?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12499677
 ] 

Julian Neil commented on HADOOP-573:
------------------------------------

I've been seeing the same problem, and other checksum problems. I am somewhat 
sceptical of the suggestion that is a memory hardware issue, but to be thorough 
I tried replacing my memory.  The errors continued. If there is any additional 
information I can provide to help track the probem down, please let me know.  
Running on a single Windows Server 2003 (with cygwin) as both namenode and 
datanode.
Strangely, some large map/reduce jobs never get checksum errors in the maps or 
reduces, but one particular job always does. 
 
In addition I have been getting many lost map outputs due to checksum errors.  
The error usually disappears when the task is retried:

Map output lost, rescheduling: getMapOutput(task_0008_m_000007_0,0) failed :
org.apache.hadoop.fs.ChecksumException: Checksum error: 
/tmp/hadoop-sshd_server/mapred/local/task_0008_m_000007_0/file.out at 60215808
        at 
org.apache.hadoop.fs.ChecksumFileSystem$FSInputChecker.verifySum(ChecksumFileSystem.java:258)
        at 
org.apache.hadoop.fs.ChecksumFileSystem$FSInputChecker.readBuffer(ChecksumFileSystem.java:211)
        at 
org.apache.hadoop.fs.ChecksumFileSystem$FSInputChecker.read(ChecksumFileSystem.java:167)
        at 
org.apache.hadoop.fs.FSDataInputStream$PositionCache.read(FSDataInputStream.java:41)
        at java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
        at java.io.BufferedInputStream.read(BufferedInputStream.java:317)
        at java.io.DataInputStream.read(DataInputStream.java:132)
        at 
org.apache.hadoop.mapred.TaskTracker$MapOutputServlet.doGet(TaskTracker.java:1674)
        at javax.servlet.http.HttpServlet.service(HttpServlet.java:689)
        at javax.servlet.http.HttpServlet.service(HttpServlet.java:802)
        at 
org.mortbay.jetty.servlet.ServletHolder.handle(ServletHolder.java:427)
        at 
org.mortbay.jetty.servlet.WebApplicationHandler.dispatch(WebApplicationHandler.java:475)
        at 
org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:567)
        at org.mortbay.http.HttpContext.handle(HttpContext.java:1565)
        at 
org.mortbay.jetty.servlet.WebApplicationContext.handle(WebApplicationContext.java:635)
        at org.mortbay.http.HttpContext.handle(HttpContext.java:1517)
        at org.mortbay.http.HttpServer.service(HttpServer.java:954)
        at org.mortbay.http.HttpConnection.service(HttpConnection.java:814)
        at org.mortbay.http.HttpConnection.handleNext(HttpConnection.java:981)
        at org.mortbay.http.HttpConnection.handle(HttpConnection.java:831)
        at 
org.mortbay.http.SocketListener.handleConnection(SocketListener.java:244)
        at org.mortbay.util.ThreadedServer.handle(ThreadedServer.java:357)
        at org.mortbay.util.ThreadPool$PoolThread.run(ThreadPool.java:534)

I'm alse getting errors in final output of the previous map/reduce job which is 
fed in as input to the next job.  These errors do not disappear when the map 
task retries:

org.apache.hadoop.fs.ChecksumException: Checksum error: 
hdfs://xxx.xxx.xxx:9900/aa/datamining/deviations_part-00002_step-00001/part-00000
 at 13781504
        at 
org.apache.hadoop.fs.ChecksumFileSystem$FSInputChecker.verifySum(ChecksumFileSystem.java:258)
        at 
org.apache.hadoop.fs.ChecksumFileSystem$FSInputChecker.readBuffer(ChecksumFileSystem.java:211)
        at 
org.apache.hadoop.fs.ChecksumFileSystem$FSInputChecker.read(ChecksumFileSystem.java:167)
        at 
org.apache.hadoop.fs.FSDataInputStream$PositionCache.read(FSDataInputStream.java:41)
        at java.io.BufferedInputStream.fill(BufferedInputStream.java:218)
        at java.io.BufferedInputStream.read(BufferedInputStream.java:237)
        at 
org.apache.hadoop.fs.FSDataInputStream$Buffer.read(FSDataInputStream.java:93)
        at java.io.DataInputStream.readInt(DataInputStream.java:372)
        at org.apache.hadoop.io.SequenceFile$Reader.next(SequenceFile.java:1523)
        at org.apache.hadoop.io.SequenceFile$Reader.next(SequenceFile.java:1436)
        at org.apache.hadoop.io.SequenceFile$Reader.next(SequenceFile.java:1482)
        at 
org.apache.hadoop.mapred.SequenceFileRecordReader.next(SequenceFileRecordReader.java:73)
        at org.apache.hadoop.mapred.MapTask$1.next(MapTask.java:157)
        at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:46)
        at org.apache.hadoop.mapred.MapTask.run(MapTask.java:175)
        at 
org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:1445)


> Checksum error during sorting in reducer
> ----------------------------------------
>
>                 Key: HADOOP-573
>                 URL: https://issues.apache.org/jira/browse/HADOOP-573
>             Project: Hadoop
>          Issue Type: Bug
>          Components: mapred
>            Reporter: Runping Qi
>            Assignee: Owen O'Malley
>
> Many reduce tasks got killed due to checksum error. The strange thing is that 
> the file was generated by the sort function, and was on a local disk. Here is 
> the stack: 
> Checksum error:  ../task_0011_r_000140_0/all.2.1 at 5342920704
>       at 
> org.apache.hadoop.fs.FSDataInputStream$Checker.verifySum(FSDataInputStream.java:134)
>       at 
> org.apache.hadoop.fs.FSDataInputStream$Checker.read(FSDataInputStream.java:110)
>       at 
> org.apache.hadoop.fs.FSDataInputStream$PositionCache.read(FSDataInputStream.java:170)
>       at java.io.BufferedInputStream.fill(BufferedInputStream.java:218)
>       at java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
>       at java.io.BufferedInputStream.read(BufferedInputStream.java:313)
>       at java.io.DataInputStream.readFully(DataInputStream.java:176)
>       at 
> org.apache.hadoop.io.DataOutputBuffer$Buffer.write(DataOutputBuffer.java:55)
>       at org.apache.hadoop.io.DataOutputBuffer.write(DataOutputBuffer.java:89)
>       at 
> org.apache.hadoop.io.SequenceFile$Reader.readBuffer(SequenceFile.java:1061)
>       at 
> org.apache.hadoop.io.SequenceFile$Reader.seekToCurrentValue(SequenceFile.java:1126)
>       at 
> org.apache.hadoop.io.SequenceFile$Reader.nextRaw(SequenceFile.java:1354)
>       at 
> org.apache.hadoop.io.SequenceFile$Sorter$MergeStream.next(SequenceFile.java:1880)
>       at 
> org.apache.hadoop.io.SequenceFile$Sorter$MergeQueue.merge(SequenceFile.java:1938)
>       at 
> org.apache.hadoop.io.SequenceFile$Sorter$MergePass.run(SequenceFile.java:1802)
>       at 
> org.apache.hadoop.io.SequenceFile$Sorter.mergePass(SequenceFile.java:1749)
>       at org.apache.hadoop.io.SequenceFile$Sorter.sort(SequenceFile.java:1494)
>       at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:240)
>       at 
> org.apache.hadoop.mapred.TaskTracker$Child.main(TaskTracker.java:1066)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to