Ideally what should be the block size in hadoop?
What happen when namenode enters in safemode in hadoop?
What is partitioning?
Can we change the file cached by distributed cache
Why would nosql be better than using a sql database? And how much better is it?
Why we use intwritable instead of int? Why we use longwritable instead of long?
What happens if you don?t override the mapper methods and keep them as it is?
What are the side data distribution techniques?
Can we deploy job tracker other than name node?
Did you ever ran into a lop sided job that resulted in out of memory error, if yes then how did you handled it ?
How many maps are there in a particular job?
Can I set the number of reducers to zero?
Doesn’t google have its very own version of dfs?
What are the restriction to the key and value class ?
If we want to copy 10 blocks from one machine to another, but another machine can copy only 8.5 blocks, can the blocks be broken at the time of replication?