Is it necessary to install spark on all the nodes of a YARN cluster while running Apache Spark on YARN ?
Answer / Swadesh Kumar Niranjan
When using Apache Spark with YARN (Yet Another Resource Negotiator), you only need to install Spark's dependency libraries on the worker nodes (nodes that run executors). The master node, which runs the application master, must have Spark installed. It is not necessary to install Spark itself on all worker nodes.
| Is This Answer Correct ? | 0 Yes | 0 No |
Are spark dataframes distributed?
What is rdd map?
How do you set up a spark?
How many types of Transformation are there?
Explain the level of parallelism in Spark Streaming? Also, describe its need.
Is spark distributed computing?
Describe Spark SQL?
Explain distnct(),union(),intersection() and substract() transformation in Spark?
Explain the terms Spark Partitions and Partitioners?
What is in memory processing in spark?
What is spark in python?
How do I clear my spark cache?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)