What are the various levels of persistence in Apache Spark?
Answer / Kailash Chandra Singh
Apache Spark provides three levels of RDD persistence: (1) StorageLevel.MEMORY_ONLY, which stores data only in memory and does not write to disk unless explicitly asked to do so. (2) StorageLevel.MEMORY_AND_DISK_SER, which stores data in memory as well as on disk using serialized format. (3) StorageLevel.MEMORY_ONLY_2, which is similar to MEMORY_AND_DISK_SER but uses a more efficient version of serialization.
| Is This Answer Correct ? | 0 Yes | 0 No |
What is the difference between rdd and dataframe in spark?
How does yarn work with spark?
Please enumerate the various components of the Spark Ecosystem.
What is a spark standalone cluster?
Is spark sql a database?
Explain reduceByKey() Spark operation?
How is rdd distributed?
Define paired RDD in Apache Spark?
How do I check my spark status?
Which language is best for spark?
Which is better scala or python for spark?
How do you stop a spark?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)