How data is spilt in Hadoop?
how Hadoop is different from other data processing tools?
what is storage and compute nodes?
In which kind of scenarios MapReduce jobs will be more useful than PIG in Hadoop?
Can there be no Reducer?
What is Data Locality in Hadoop?
What are the four basic parameters of a reducer?
How many Reducers run for a MapReduce job in Hadoop?
What is identity mapper and reducer? In which cases can we use them?
What is mapper in map reduce?
What is heartbeat in hdfs? Explain.
What is a combiner and where you should use it?
What is Counter in MapReduce?