Usually large file in HDFS is split into bulks and store in different DataNodes. A map task is assigned to deal with that bulk, I wonder what if the Structured data(i.e a word) was split into two bulks? How MapReduce and HDFS deal with this?
Thanks! Donal