To ensure data I/O integrity, hadoop uses CRC 32 mechanism to generate
checksum for the data stored on hdfs . But suppose I have a data node machine
that does not have ecc(error correcting code) type of memory, So will hadoop
hdfs will be able to generate checksum for data blocks when
While the HDFS functionality of computing, storing and validating
checksums for block files does not specifically _require_ ECC, you do
_want_ ECC to avoid frequent checksum failures.
This is noted in Tom's book as well, in the chapter that discusses
setting up your own cluster:
ECC memory is