HADOOP-11601 tightens up the filesystem spec by saying "if len(file) > 0, 
getFileStatus().getBlockSize() > 0"

this is to stop filesystems (most recently s3a) returning 0 as a block size, 
which then kills any analytics work that tries to partition the workload by 
blocksize.

I'm currently changing the markdown text to say

MUST be >0 for a file size >0
MAY be 0 for a file of size==0.

+ the relevant tests to check this.

There's one thing I do want to understand from HDFS first: what about small 
files.? That is: what does HDFS return as a blocksize if a file is smaller than 
its block size?

-Steve

Reply via email to