I've been reading in the "Hadoop Definitive guide" by Tom White
about the block emptiness when a file is not large enough to occupy the full size of the block.
From the statement (cite from the book)
"Unlike a filesystem for a single disk, a file in HDFS that is smaller than a single block does not occupy a full block’s worth of underlying storage." I can understand the physical space left from the initial block size will be free. My question is can the underlying operating reuse/write this remained free space?
I'll look forward for your answers.