Replication causes data redundancy then why is pursued in hdfs?
Answer / Lalita Kumari
Replication is pursued to ensure high availability and fault tolerance of data in HDFS. In case a node fails, the replicated data can be used from other nodes.
| Is This Answer Correct ? | 0 Yes | 0 No |
What is a block in Hadoop HDFS? What should be the block size to get optimum performance from the Hadoop cluster?
Can we change the document present in hdfs?
Why is block size large in Hadoop?
Explain how are file systems checked in hdfs?
Can multiple clients write into an HDFS file concurrently in hadoop?
Explain the difference between an hdfs block and input split?
Define hadoop archives? What is the command for archiving a group of files in hdfs.
If data is present in HDFS and RF is defined, then how can we change Replication Factor?
While processing data from hdfs, does it execute code near data?
Will various customers write into an hdfs record simultaneously?
What is the problem in having lots of small files in hdfs?
How data or file is written into Hadoop HDFS?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)