Hi all: The absolute time required to check and repair a file system is increasing because disk capacities are growing faster than disk bandwidth and seek time remains almost unchanged.
At the same time, file system repair is becoming more common, because the perbit error rate of disks is not dropping as fast as the number of bits per disk is growing, resulting in more errors per disk. With existing file systems, a single corrupted metadata block requires the entire file system to be unmounted, checked, and repaired a process that takes hours or days to complete, during which time the data is completely unavailable. The resulting fsck time crunch is already making file systems only a few terabytes in size impractical to administrate. The paper propose chunkfs, which divides on disk file system data into small, individually repairable faultisolation domains while preserving normal file system semantics. Take a look at the complete paper at - http://infohost.nmt.edu/~val/review/chunkfs.pdf thanks Saifi.

