How to create RDD?
Answer / Naveen Singh Kuldeep
RDDs can be created in Apache Spark using built-in functions like textFile(filePath), wholeTextFiles(filePath), parallelize(iterable), and fromJavaRDD/fromJavaCollectionRDD. For example, val rdd = sc.textFile("input.txt").
| Is This Answer Correct ? | 0 Yes | 0 No |
What are the great features of spark sql?
What is the biggest shortcoming of Spark?
What is the difference between cache and persist in spark?
Why is there a need for broadcast variables when working with Apache Spark?
What is SparkSession in Apache Spark? Why is it needed?
How to save RDD?
Is there any benefit of learning mapreduce if spark is better than mapreduce?
Can you run spark without hadoop?
What is lambda in spark?
Can you explain spark mllib?
What are the features of RDD, that makes RDD an important abstraction of Spark?
What do you mean by Speculative execution in Apache Spark?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)