HDFS is used for applications with large data sets, not why Many small files?
Answer Posted / Jaiprakash
HDFS is designed for storing large data sets. One of the challenges with many small files is that they increase the number of blocks in HDFS, which can lead to increased metadata overhead and slower operations. However, there are techniques like combining multiple small files into larger ones (e.g., using the DistCP utility) to mitigate these issues.
| Is This Answer Correct ? | 0 Yes | 0 No |
Post New Answer View All Answers
No New Questions to Answer in this Category !! You can
Post New Questions
Answer Questions in Different Category