Usually large file in HDFS is split into bulks and store in different
DataNodes.
A map task is assigned to deal with that bulk, I wonder what if the
Structured data(i.e a word) was split into two bulks?
How MapReduce and HDFS deal with this?

Thanks!
Donal

Reply via email to