1.

Explain About The Common Workflow Of A Spark Program

Answer»
  • The foremost step in a Spark program INVOLVES creating input RDD's from EXTERNAL data.
  • Use various RDD TRANSFORMATIONS like filter() to create new transformed RDD's based on the BUSINESS logic.
  • persist() any intermediate RDD's which MIGHT have to be reused in future.
  • Launch various RDD actions() like first(), count() to begin parallel computation , which will then be optimized and executed by Spark.



Discussion

No Comment Found