When we create an rdd, does it bring the data and load it into the memory?
Answer / Apurav Garg
No, creating an RDD in Apache Spark does not automatically load all the data into memory. By default, only part of the data is loaded into memory as needed to perform transformations or actions. If more data is required during computation, additional data can be fetched from storage.
| Is This Answer Correct ? | 0 Yes | 0 No |
What is spark driver application?
Can you explain benefits of spark over mapreduce?
What is data pipeline in spark?
Define Partitions?
What are the roles and responsibilities of worker nodes in the Apache Spark cluster? Is Worker Node in Spark is same as Slave Node?
What is sc parallelize?
What are the ways in which Apache Spark handles accumulated Metadata?
Define the term ‘sparse vector.’
What are the ways to run spark over hadoop?
How will you connect Apache Spark with Apache Mesos?
Explain how can spark be connected to apache mesos?
How do you parse data in xml? Which kind of class do you use with java to parse data?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)