What is aggregatebykey spark?
Answer / Shadab Faisal
"AggregateByKey is a Spark function that allows you to perform a custom aggregation operation on each key of an RDD, combining values for the same key. It returns a new RDD with the result of the aggregation. For example, you can use AggregateByKey to count the occurrences of each word in a text file like so: `textRDD.flatMapValues(_.split(" ")).aggregateByKey(0)(_+1, _+_)`"n
| Is This Answer Correct ? | 0 Yes | 0 No |
What is the use of map transformation?
What are the ways to run spark over hadoop?
Explain join() operation in Apache Spark?
What is driver and executor in spark?
What are the disadvantages of using Spark?
Is rdd type safe?
Does spark sql use hive?
Why spark is faster than hive?
Explain distnct(),union(),intersection() and substract() transformation in Spark?
What are the great features of spark sql?
Can spark be used without hadoop?
How can you trigger automatic clean-ups in Spark to handle accumulated metadata?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)