Illustrate some demerits of using Spark.
Answer / Renu Grover
Demerits of using Spark include: High resource consumption, especially for small datasets; increased overhead compared to MapReduce due to additional functionality like in-memory caching and iterative algorithms; potential scalability issues when dealing with extremely large datasets; and potential complexity for configuring and managing the Spark cluster.
| Is This Answer Correct ? | 0 Yes | 0 No |
What is shuffle read and shuffle write in spark?
What is number of executors in spark?
What is hive on spark?
Define the level of parallelism and its need in spark streaming?
Explain Catalyst framework?
Does Spark provide the storage layer too?
What is the difference between DAG and Lineage?
What is the difference between reducebykey and groupbykey?
Where is spark rdd?
What is lineage graph?
What is salting in spark?
Explain leftOuterJoin() and rightOuterJoin() operation in Apache Spark?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)