Apache Spark Pertanyaan dan Jawaban Wawancara
Question: What is the purpose of the SparkContext in Apache Spark?Answer: SparkContext is the entry point for Spark functionality and represents the connection to the Spark cluster. It coordinates the execution of operations on the cluster.Example:
|
Simpan untuk Revisi
Bookmark item ini, tandai sebagai sulit, atau masukkan ke dalam set revisi.
Masuk untuk menyimpan bookmark, pertanyaan sulit, dan set revisi.
Apakah ini membantu? Ya Tidak
Most helpful rated by users:
- What is the purpose of the Spark SQL module?
- Explain the difference between narrow and wide transformations in Spark.