Do you need to install spark on all nodes of yarn cluster?
Answer / Desh Deepak
When running a Spark application on a YARN cluster, it is not necessary to install Spark on all nodes. Instead, only the resource manager and worker nodes with required resources (memory and CPU) should have Spark installed. The rest of the worker nodes can continue to run other YARN applications.
| Is This Answer Correct ? | 0 Yes | 0 No |
What are the disadvantages of using Spark?
Does spark require hdfs?
How many ways we can create rdd in spark?
How does yarn work with spark?
Is there any benefit of learning mapreduce if spark is better than mapreduce?
How do we represent data in Spark?
Can spark be used without hadoop?
What is shuffle spill in spark?
Is spark and hadoop same?
What is data pipeline in spark?
When creating an RDD, what goes on internally?
What is the difference between client mode and cluster mode in spark?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)