Why is Reading done in parallel and writing is not in HDFS?
Answer / Lalita Kumari
Reading is done in parallel in HDFS because it's more efficient for the system to read multiple blocks simultaneously. Writing, however, requires exclusive access to the file, so it cannot be performed in parallel.
| Is This Answer Correct ? | 0 Yes | 0 No |
If the source data gets updated every now and then, how will you synchronize the data in hdfs that is imported by sqoop?
What is the throughput? How does hdfs give great throughput?
Does the HDFS go wrong? If so, how?
Can you explain about the indexing process in hdfs?
How does hdfs provides good throughput?
How are file systems checked in hdfs?
Why is HDFS only suitable for large data sets and not the correct tool to use for many small files?
Why is block size large in Hadoop?
How does HDFS ensure Data Integrity of data blocks stored in HDFS?
How is hdfs block size different from traditional file system block size?
How HDFS client divide the file into the block while storing inside HDFS?
What is the difference between an hdfs block and input split?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)