I know that HDFS stores data using the regular linux file system in the data nodes. My HDFS block size is 128 MB
. Lets say that I have 10 GB
of disk space in my hadoop cluster that means, HDFS initially has 80 blocks
as available storage.
If I create a small file of say 12.8 MB
, #available HDFS blocks will become 79. What happens if I create another small file of 12.8 MB
? Will the #availbale blocks stay at 79 or will it come down to 78? In the former case, HDFS basically recalculates the #available blocks after each block allocation based on the available free disk space so, #available blocks will become 78 only after more than 128 MB of disk space is consumed. Please clarify.