>Why would you want to use a block size of > 2GB?
For keeping a maps input split in a single block~

On Tue, Oct 19, 2010 at 9:07 AM, Michael Segel <michael_se...@hotmail.com>wrote:

>
> Ok, I'll bite.
> Why would you want to use a block size of > 2GB?
>
>
>
> > Date: Mon, 18 Oct 2010 21:33:34 +1100
> > Subject: BUG: Anyone use block size more than 2GB before?
> > From: eltonsky9...@gmail.com
> > To: common-user@hadoop.apache.org
> >
> > Hello,
> >
> > In
> >
> hdfs.org.apache.hadoop.hdfs.DFSClient<eclipse-javadoc:%E2%98%82=HadoopSrcCode/src%3Chdfs.org.apache.hadoop.hdfs%7BDFSClient.java%E2%98%83DFSClient>
> >
> .DFSOutputStream<eclipse-javadoc:%E2%98%82=HadoopSrcCode/src%3Chdfs.org.apache.hadoop.hdfs%7BDFSClient.java%E2%98%83DFSClient%E2%98%83DFSOutputStream>.writeChunk(byte[]
> > b, int offset, int len, byte[] checksum)
> > The second last line:
> >
> > int psize = Math.min((int)(blockSize-bytesCurBlock), writePacketSize);
> >
> > When I use blockSize  bigger than 2GB, which is out of the boundary of
> > integer something weird would happen. For example, for a 3GB block it
> will
> > create more than 2Million packets.
> >
> > Anyone noticed this before?
> >
> > Elton
>
>

Reply via email to