[ http://issues.apache.org/jira/browse/HADOOP-646?page=all ]
Milind Bhandarkar updated HADOOP-646:
-------------------------------------
Attachment: edits.patch
Patch for christian to test his large edits file loading.
> name node server does not load large (> 2^31 bytes) edits file
> --------------------------------------------------------------
>
> Key: HADOOP-646
> URL: http://issues.apache.org/jira/browse/HADOOP-646
> Project: Hadoop
> Issue Type: Bug
> Components: dfs
> Affects Versions: 0.7.1
> Reporter: Christian Kunz
> Priority: Critical
> Attachments: edits.patch
>
>
> FileInputStream.available() returns negative values when reading a large file
> (> 2^31 bytes) -- this is a known (unresolved) java bug:
> http://bugs.sun.com/bugdatabase/view_bug.do?bug_id=6402006
> Consequence: a large edits file is not loaded and deleted without any
> warnings. The system reverts back to the old fsimage.
> This happens in jdk1.6 as well, i.e. the bug has not yet been fixed.
> In addition, when finally I was able to load my big cron-backed-up edits file
> (6.5 GB) with a kludgy work-around, the blocks did not exist anymore in the
> data node servers, probably deleted from the previous attempts when the name
> node server did not know about the changed situation.
> Moral till this is fixed or worked-around: don't wait too long to restart the
> name node server. Otherwise this is a way to lose the entire dfs.
--
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators:
http://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see: http://www.atlassian.com/software/jira