Why is rdd immutable?
Answer / Ankur Mittal
RDD (Resilient Distributed Datasets) in Apache Spark are designed to be immutable for fault tolerance. Once an RDD is created, its data cannot be altered directly to ensure that when a task fails, the computation can be re-executed on another node with consistent data.
| Is This Answer Correct ? | 0 Yes | 0 No |
What does repartition do in spark?
What is spark written?
What languages support spark?
How do you stop a spark?
What do you understand by Lazy Evaluation?
What are the features of apache spark?
What is Directed Acyclic Graph in Apache Spark?
Is there a module to implement sql in spark?
Is apache spark going to replace hadoop?
Explain about trformations and actions in the context of rdds?
How does spark work with python?
What according to you is a common mistake apache spark developers make when using spark ?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)