How do sparks work?
Answer / Ankush Panwar
Spark processes data in resilient distributed datasets (RDDs), which are collections of data that can be stored and manipulated across a Spark cluster. Each RDD is partitioned into smaller chunks called tasks, which are executed by worker nodes. Spark uses lineage to keep track of the dependencies between different operations and optimize computations.
| Is This Answer Correct ? | 0 Yes | 0 No |
How spark is faster than hadoop?
Is hadoop mandatory for spark?
Explain the various Transformation on Apache Spark RDD like distinct(), union(), intersection(), and subtract()?
What does map transformation do? Provide an example.
How spark works on hadoop?
What is the difference between hive and spark?
What is spark lineage?
What is a spark rdd?
Is apache spark a framework?
What is pagerank in graphx?
By Default, how many partitions are created in RDD in Apache Spark?
What is project tungsten in spark?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)