The default data block size of HDFS/hadoop is 64MB.
The block size in disk is generally 4KB.
What does 64MB block size mean? ->Does it mean that the smallest unit of read fr
If block size was set to less than 64, there would be a huge number of blocks throughout the cluster, which causes NameNode to manage an enormous amount of metadata.
Since we need a Mapper for each block, there would be a lot of Mappers, each processing a piece bit of data, which isn't efficient.