Why do we use HDFS for applications having large data sets and not when there are lot of small files?
1 2105
What do you mean by the high availability of a namenode?
Why is block size set to 128 MB in Hadoop HDFS?
If you run hive as a server, what are the available mechanism for connecting it from application?
What are the areas where ambari helps the system administrators to do?
If a particular file is 50 mb, will the hdfs block still consume 64 mb as the default size?
Can we use windows for hadoop?
How does impala achieve its performance improvements?
How does pipe operation writes the result to standard output in Apache Spark?
How will you make changes to the default configuration files?
What are the problems with Hadoop 1.0?
Is spark secure?
What is executor cores in spark?
What is rdd map?
What is streaming?
What operations does rdd support?