Answer Posted / Neeraj P Singh
"Resilient Distributed Datasets (RDD) are the basic data structure of Apache Spark. An RDD is an immutable distributed collection of objects that can be processed in parallel across a cluster. For example, you can create an RDD from a list of integers using the `spark.sparkContext.parallelize()` method like so: `val rdd = sparkContext.parallelize(Array(1, 2, 3, 4))`."n
| Is This Answer Correct ? | 0 Yes | 0 No |
Post New Answer View All Answers