[ http://issues.apache.org/jira/browse/HADOOP-643?page=all ]
Raghu Angadi resolved HADOOP-643.
---------------------------------
Resolution: Duplicate
Assignee: Raghu Angadi
Duplicate of HADOOP-604.
> failure closing block of file
> -----------------------------
>
> Key: HADOOP-643
> URL: http://issues.apache.org/jira/browse/HADOOP-643
> Project: Hadoop
> Issue Type: Bug
> Components: dfs
> Affects Versions: 0.7.2
> Reporter: Johan Oskarson
> Assigned To: Raghu Angadi
> Priority: Critical
>
> I've been getting "failure closing block of file" on random files.
> Both datanode and tasktracker running on node7. No problems with pinging.
> Guess it got stuck after the NPE in DataNode.
> Job cannot start because of:
> java.io.IOException: failure closing block of file
> /home/hadoop/mapred/system/submit_99u9cd/.job.jar.crc to node node7:50010
> at
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.internalClose(DFSClient.java:1199)
> at
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.endBlock(DFSClient.java:1163)
> at
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.close(DFSClient.java:1241)
> at java.io.FilterOutputStream.close(FilterOutputStream.java:143)
> at java.io.FilterOutputStream.close(FilterOutputStream.java:143)
> at java.io.FilterOutputStream.close(FilterOutputStream.java:143)
> at
> org.apache.hadoop.fs.FSDataOutputStream$Summer.close(FSDataOutputStream.java:96)
> at java.io.FilterOutputStream.close(FilterOutputStream.java:143)
> at java.io.FilterOutputStream.close(FilterOutputStream.java:143)
> at java.io.FilterOutputStream.close(FilterOutputStream.java:143)
> at org.apache.hadoop.fs.FileUtil.copyContent(FileUtil.java:205)
> at org.apache.hadoop.fs.FileUtil.copyContent(FileUtil.java:190)
> at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:77)
> at
> org.apache.hadoop.dfs.DistributedFileSystem.copyFromLocalFile(DistributedFileSystem.java:186)
> at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:289)
> at org.apache.hadoop.mapred.jobcontrol.Job.submit(Job.java:314)
> at
> org.apache.hadoop.mapred.jobcontrol.JobControl.startReadyJobs(JobControl.java:248)
> at
> org.apache.hadoop.mapred.jobcontrol.JobControl.run(JobControl.java:280)
> at java.lang.Thread.run(Thread.java:595)
> Caused by: java.net.SocketTimeoutException: Read timed out
> at java.net.SocketInputStream.socketRead0(Native Method)
> at java.net.SocketInputStream.read(SocketInputStream.java:129)
> at java.io.BufferedInputStream.fill(BufferedInputStream.java:218)
> at java.io.BufferedInputStream.read1(BufferedInputStream.java:256)
> at java.io.BufferedInputStream.read(BufferedInputStream.java:313)
> at java.io.DataInputStream.readFully(DataInputStream.java:176)
> at java.io.DataInputStream.readLong(DataInputStream.java:380)
> at
> org.apache.hadoop.dfs.DFSClient$DFSOutputStream.internalClose(DFSClient.java:1193)
> Cxception in datanode.out on node7:
> Exception in thread "[EMAIL PROTECTED]" java.lang.NullPointerException
> at
> org.apache.hadoop.dfs.FSDataset$FSDir.checkDirTree(FSDataset.java:162)
> at
> org.apache.hadoop.dfs.FSDataset$FSDir.checkDirTree(FSDataset.java:162)
> at
> org.apache.hadoop.dfs.FSDataset$FSVolume.checkDirs(FSDataset.java:238)
> at
> org.apache.hadoop.dfs.FSDataset$FSVolumeSet.checkDirs(FSDataset.java:326)
> at org.apache.hadoop.dfs.FSDataset.checkDataDir(FSDataset.java:522)
> at
> org.apache.hadoop.dfs.DataNode$DataXceiveServer.run(DataNode.java:480)
> at java.lang.Thread.run(Thread.java:595)
--
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators:
http://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see: http://www.atlassian.com/software/jira