Each block of each file is scattered on (currently) three random data nodes, not related to the previous block placement. So no, no limits on file size until you reach the FS limits, which are reasonably high and growing (probably a couple 100 TB currently).
-----Original Message----- From: Dennis Kubes [mailto:[EMAIL PROTECTED] Sent: Sunday, March 26, 2006 12:12 PM To: [email protected] Subject: Hadoop File Capacity For the Hadoop filesystem, I know that it is basically unlimited in terms of storage because one can always add new hardware, but it is unlimited in terms of a single file? What I mean by this is if I store a file /user/dir/a.index and this file has say 100 blocks in it where there is only enough space on any server for 10 blocks; will the Hadoop filesystem store and replicate different blocks on different servers and give the client a single file view or does a whole file have to be stored and replicated across machines. Dennis
