Perhaps we should implement a system by which removed blocks linger
on the data nodes for a long period of time (a week?) (or until the
space is needed fifo)?
Ditto with the meta data? That would make archeology around failures
easier.
On Mar 23, 2006, at 4:56 PM, Doug Cutting (JIRA) wrote:
[ http://issues.apache.org/jira/browse/HADOOP-101?
page=comments#action_12371659 ]
Doug Cutting commented on HADOOP-101:
-------------------------------------
I like that this does not use anything more than the client API to
check the server. That keeps the server core lean and mean. The
use of RPC's effectively restricts the impact of the scan on the FS.
A datanode operation that streams through a block without
transferring it over the wire won't correctly check checksums using
our existing mechanism. To check file content we could instead
simply implement a map-reduce job that streams through all the
files in the fs. This would not take much code: nothing additional
in the core. MapReduce should handle the locality, so that most
data shouldn't go over the wire.
BTW, blocks not used by any file are not known to the name node,
are they? When they're reported by a datanode the datanode is told
to remove them.
DFSck - fsck-like utility for checking DFS volumes
--------------------------------------------------
Key: HADOOP-101
URL: http://issues.apache.org/jira/browse/HADOOP-101
Project: Hadoop
Type: New Feature
Components: dfs
Versions: 0.2
Reporter: Andrzej Bialecki
Assignee: Andrzej Bialecki
Attachments: DFSck.java
This is a utility to check health status of a DFS volume, and
collect some additional statistics.
--
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the
administrators:
http://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see:
http://www.atlassian.com/software/jira