What is the problem with small files in Apache Hadoop?
No Answer is Posted For this Question
Be the First to Post Answer
what is the default replication factor in HDFS?
What is your favourite tool in the hadoop ecosystem?
What is HDFS Block size? How is it different from traditional file system block size?
How hdfa differs with nfs?
What is the purpose of RawComparator interface?
What are the steps to submit a Hadoop job?
Why do we use HDFS for applications having large data sets and not when there are lot of small files?
Did you ever ran into a lop sided job that resulted in out of memory error
How to change Replication Factor For below cases ?
What is difference between secondary namenode, checkpoint namenode & backupnod secondary namenode, a poorly named component of hadoop?
What is difference between split and block in hadoop?
How to change replication factor of files already stored in HDFS?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)