Explain the run-time architecture of Spark?
Answer / Anjali Nayak
"Spark's runtime architecture consists of a Driver Program (Java or Scala code) that submits tasks to the Cluster Manager (YARN, Mesos, or Standalone). The Cluster Manager then allocates resources for executors. Each Executor runs multiple tasks concurrently and manages memory and storage through RDDs (Resilient Distributed Datasets). Spark uses DAG (Directed Acyclic Graph) scheduler to optimize task execution."
| Is This Answer Correct ? | 0 Yes | 0 No |
Can you define parquet file?
Explain reduceByKey() Spark operation?
What is meant by rdd lazy evaluation?
How do we represent data in Spark?
Explain about transformations and actions in the context of RDDs.
Is there any benefit of learning mapreduce if spark is better than mapreduce?
What is difference between spark and mapreduce?
What is a hive on spark?
Can you run spark without hadoop?
What rdd stands for?
How do you integrate spark and hive?
What is executor spark?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)