Ok, I'll bite. Why would you want to use a block size of > 2GB?
> Date: Mon, 18 Oct 2010 21:33:34 +1100 > Subject: BUG: Anyone use block size more than 2GB before? > From: [email protected] > To: [email protected] > > Hello, > > In > hdfs.org.apache.hadoop.hdfs.DFSClient<eclipse-javadoc:%E2%98%82=HadoopSrcCode/src%3Chdfs.org.apache.hadoop.hdfs%7BDFSClient.java%E2%98%83DFSClient> > .DFSOutputStream<eclipse-javadoc:%E2%98%82=HadoopSrcCode/src%3Chdfs.org.apache.hadoop.hdfs%7BDFSClient.java%E2%98%83DFSClient%E2%98%83DFSOutputStream>.writeChunk(byte[] > b, int offset, int len, byte[] checksum) > The second last line: > > int psize = Math.min((int)(blockSize-bytesCurBlock), writePacketSize); > > When I use blockSize bigger than 2GB, which is out of the boundary of > integer something weird would happen. For example, for a 3GB block it will > create more than 2Million packets. > > Anyone noticed this before? > > Elton
