s to <= 64MB (what i did
> and recommend)
> or reconfigure your hadoop.
>
>
> dfs.block.size
> 67108864
> The default block size for new files.
>
>
> do something like
> cat * | rotatelogs ./merged/m 64M
> it will merge and chop the data fo
hi all -
can anyone comment on the performance cost of merging many small files into
an increasingly large MapFile ? will that cost be dependent on the size of
the larger MapFile (since I have to rewrite it) or is there a built-in
strategy to split it into smaller parts, affecting only those which