Why do we use HDFS for applications having large data sets and not when there are lot of small files?
1 2777Post New Apache Hadoop Questions
What is rack-aware replica placement policy?
What are the modules that constitute the Apache Hadoop 2.0 framework?
Did you ever ran into a lop sided job that resulted in out of memory error
How can we check whether namenode is working or not?
What is the communication channel between client and namenode/datanode?
Explain why the name ‘hadoop’?
On what basis data will be stored on a rack?
Who is a 'user' in HDFS?
What are the main components of a Hadoop Application?
what should be the ideal replication factor in hadoop?
What Mapper does?
What is HDFS ? How it is different from traditional file systems?
What is the difference between Gen1 and Gen2 Hadoop with regards to the Namenode?
How Mapper is instantiated in a running job?
What is Derby database?