What are the features of spark rdd?
Answer / Sushil Kumar Yadav
RDD (Resilient Distributed Datasets) in Spark is a key concept. It is an immutable distributed collection of objects, each with a partition. RDDs can be transformed and acted upon in parallel across the cluster. They also have fault tolerance mechanisms to handle node failures.
| Is This Answer Correct ? | 0 Yes | 0 No |
What is Directed Acyclic Graph in Apache Spark?
What is spark checkpointing?
What is external shuffle service in spark?
Does spark run hadoop?
How spark works on hadoop?
What are the common transformations in apache spark?
Define Spark-SQL?
Explain the repartition() operation in Spark?
Is spark difficult to learn?
Does spark replace hadoop?
Does spark sql use hive?
Describe Partition and Partitioner in Apache Spark?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)