Answer Posted / Shweta Kashyap
"Apache Spark processes large datasets by dividing the data into smaller pieces called Resilient Distributed Datasets (RDDs) and distributing these pieces across a cluster of machines. Each machine runs tasks that operate on its assigned portion of the data, and results are aggregated to produce the final output. Spark also provides high-level APIs for common big data processing tasks like SQL, streaming, and machine learning."n
| Is This Answer Correct ? | 0 Yes | 0 No |
Post New Answer View All Answers