Re: Block Size

2011-09-29 Thread Uma Maheswara Rao G 72686
hi, Here is some useful info: A small file is one which is significantly smaller than the HDFS block size (default 64MB). If you’re storing small files, then you probably have lots of them (otherwise you wouldn’t turn to Hadoop), and the problem is that HDFS can’t handle lots of files. Every

Re: Block Size

2011-09-29 Thread Chris Smith
On 29 September 2011 18:39, lessonz wrote: > I'm new to Hadoop, and I'm trying to understand the implications of a 64M > block size in the HDFS. Is there a good reference that enumerates the > implications of this decision and its effects on files stored in the system > as well as map-reduce jobs?

Re: Re: block size

2011-09-20 Thread hao.wang
Hi, Joey: Thanks for your help! 2011-09-21 hao.wang 发件人: Joey Echeverria 发送时间: 2011-09-21 10:10:54 收件人: common-user 抄送: 主题: Re: block size HDFS blocks are stored as files in the underlying filesystem of your datanodes. Those files do not take a fixed amount of space, so if you

Re: block size

2011-09-20 Thread Joey Echeverria
HDFS blocks are stored as files in the underlying filesystem of your datanodes. Those files do not take a fixed amount of space, so if you store 10 MB in a file and you have 128 MB blocks, you still only use 10 MB (times 3 with default replication). However, the namenode does incur additional over