Why is transformation lazy operation in Apache Spark RDD? How is it useful?
Answer / Rahul Bajpai
Transformations in Apache Spark are lazy because they only execute when an action (like collect(), count(), or save()) is triggered. This allows for efficient task scheduling, optimization, and caching of intermediate results to improve performance for large datasets.
| Is This Answer Correct ? | 0 Yes | 0 No |
Do we need to install spark in all nodes?
What is tungsten engine in spark?
Is spark sql a database?
Explain leftOuterJoin() and rightOuterJoin() operation in Apache Spark?
Who invented spark?
What is apache spark used for?
Write the command to start and stop the spark in an interactive shell?
What are the optimization techniques in spark?
What is hive on spark?
What is sc textfile?
What is broadcast variable?
How can I improve my spark performance?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)