How does hdfs provides good throughput?
Answer / Asif Abbas Zaidi
HDFS provides good throughput by leveraging data locality, where tasks are executed on the same node as the data they are working with. This reduces network traffic and improves performance. Additionally, HDFS uses a buffer cache for frequently accessed data to minimize disk I/O.
| Is This Answer Correct ? | 0 Yes | 0 No |
What is a block in Hadoop HDFS? What should be the block size to get optimum performance from the Hadoop cluster?
Why do we need hdfs?
What do you mean by block scanner in hdfs?
Can you explain about the indexing process in hdfs?
How are file systems checked in hdfs?
Does HDFS allow a client to read a file which is already opened for writing?
Replication causes data redundancy then why is is pursued in HDFS?
What are the key features of HDFS?
What is Hadoop HDFS – Hadoop Distributed File System?
How can one set space quota in Hadoop (HDFS) directory?
Since the data is replicated thrice in hdfs, does it mean that any calculation done on one node will also be replicated on the other two?
Write the command to copy a file from linux to hdfs?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)