Why HDFS performs replication, although it results in data redundancy?
Answer / Rohit Kumar
HDFS performs replication to ensure high availability and fault tolerance. By replicating data across multiple DataNodes, if one DataNode fails, the data can still be accessed from another DataNode. Additionally, HDFS is designed for handling large-scale data processing in a distributed environment, where failures are common.
| Is This Answer Correct ? | 0 Yes | 0 No |
Define hadoop archives?
How data or a file is written into hdfs?
Why HDFS performs replication, although it results in data redundancy?
How to read file in HDFS?
Explain the hdfs architecture?
What is the procedure to create users in HDFS and how to allocate Quota to them?
Why is HDFS only suitable for large data sets and not the correct tool to use for many small files?
How to copy a file into HDFS with a different block size to that of existing block size configuration?
How is indexing done in HDFS?
Mention what is the difference between hdfs and nas?
How is hdfs block size different from traditional file system block size?
What is throughput in HDFS?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)