Answer Posted / Ankush Panwar
Spark processes data in resilient distributed datasets (RDDs), which are collections of data that can be stored and manipulated across a Spark cluster. Each RDD is partitioned into smaller chunks called tasks, which are executed by worker nodes. Spark uses lineage to keep track of the dependencies between different operations and optimize computations.
| Is This Answer Correct ? | 0 Yes | 0 No |
Post New Answer View All Answers