What are accumulators in Apache Spark?
Answer / Vivek Kumar Mishra
Accumulators in Apache Spark are variables that can be updated by tasks running on executors during the execution of an action. They provide a way to aggregate results across multiple tasks without the need for shuffling data. Accumulators can be used for various purposes, such as tracking the number of records processed or maintaining a moving average. Users can define their own accumulator classes using Scala or Java.
| Is This Answer Correct ? | 0 Yes | 0 No |
Why lazy evaluation is good in spark?
Is it necessary to learn hadoop for spark?
What is the difference between Caching and Persistence in Apache Spark?
Is there a module to implement sql in spark?
How spark is faster than hadoop?
What are the ways to run spark over hadoop?
What do you understand by receivers in Spark Streaming ?
What is big data spark?
Define Partition in Apache Spark?
Why we use parallelize in spark?
What is spark vcores?
Name a few commonly used spark ecosystems?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)