Explain in brief what is the architecture of Spark?
Answer / Avni Gupta
The architecture of Apache Spark consists of the following components: Resilient Distributed Datasets (RDD), Directed Acyclic Graph (DAG), and Spark Driver Program. RDDs are the basic building block of Spark, a distributed collection of objects. DAG is a graphical representation of a computational problem, which describes how tasks in Spark are scheduled and executed. The Spark Driver Program acts as an entry point to submit jobs to the cluster.
| Is This Answer Correct ? | 0 Yes | 0 No |
What is the difference between spark and hive?
What is a dataframe spark?
Explain Spark join() operation?
What is hdfs spark?
Do streamers make money from sparks?
What is "GraphX" in Spark?
Describe the distnct(),union(),intersection() and substract() transformation in Apache Spark RDD?
What is deploy mode in spark?
Explain the level of parallelism in Spark Streaming? Also, describe its need.
Explain the various Transformation on Apache Spark RDD like distinct(), union(), intersection(), and subtract()?
Who invented spark?
Does spark load all data in memory?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)