Why are spark transformations lazy?
Answer / Yashwant Singh Rana
Spark transformations (such as map(), filter(), and groupBy()) are lazy because they don't immediately execute the operations on the underlying data. Instead, they create new RDDs or DataFrames that represent the results of the transformation. This allows for efficient execution by batching multiple transformations together and optimizing the execution plan before actually performing the computations.
| Is This Answer Correct ? | 0 Yes | 0 No |
Explain Dsstream with reference to Apache Spark
What are the ways to create RDDs in Apache Spark? Explain.
What advantages does Spark offer over Hadoop MapReduce?
What are broadcast variables in Apache Spark? Why do we need them?
Can you explain how to minimize data transfers while working with Spark?
Explain the action count() in Spark RDD?
What do you use spark for?
What is row rdd in spark?
What does it mean by Columnar Storage Format?
Why do we use spark?
What is action, how it process data in apache spark
How do you integrate spark and hive?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)