Define the run-time architecture of Spark?
Answer / Priyanka Pandey
Apache Spark's runtime architecture is composed of three main components: (1) Spark Driver Program, (2) Executor Services, and (3) Hadoop YARN or Standalone Mode. The Spark Driver Program initiates the application, creates an RDD (Resilient Distributed Dataset), schedules tasks, and monitors the progress of the tasks. Executor Services run on worker nodes. They execute the tasks assigned by the driver program, manage memory, and communicate with other executors for resilience. Hadoop YARN or Standalone Mode manages resources for Spark applications.
| Is This Answer Correct ? | 0 Yes | 0 No |
Where are rdd stored?
Explain reduceByKey() Spark operation?
In a very huge text file, you want to just check if a particular keyword exists. How would you do this using Spark?
What are the common transformations in apache spark?
What are the various programming languages supported by Spark?
List down the languages supported by Apache Spark?
What is sparkcontext in spark?
What is dag – directed acyclic graph?
How do you process big data with spark?
Does spark sql use hive?
Describe Partition and Partitioner in Apache Spark?
What is difference between hive and spark?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)