Why do we use HDFS for applications having large data sets and not when there are lot of small files?
1 2132Post New Apache Hadoop Questions
What is DistributedCache and its purpose?
Explain a simple Map/Reduce problem.
Explain the core components of hadoop?
What are the port numbers of namenode, job tracker and task tracker?
How Mapper is instantiated in a running job?
What are sink processors?
Define fault tolerance?
What are the network requirements for using hadoop?
Is client the end user in HDFS?
How does a namenode handle the failure of the data nodes?
What is the default block size in hdfs?
What alternate way does HDFS provides to recover data in case a Namenode
Explain Erasure Coding in Apache Hadoop?
On what basis Namenode will decide which datanode to write on?
Name the various types of lists supported by bootstrap.