Summary
This concludes the chapter. We have gone through a churn prediction example using the PySpark and the Jupyiter notebook. I hope this gives you a good starting point for building your own applications. The full code and the Jupyter notebook are available on this book's GitHub page.
This was the last major chapter of this book. As a part of this book our intention was to take the users who are beginning to learn Spark on a journey where they can start from the very basics to a level where they feel comfortable with Spark as framework and also about writing their own Spark applications. We've covered some interesting topics including RDDs, DataFrames, MlLib, GraphX and also how to set up Spark in a cluster mode. Any book cannot do justice to Spark as a framework, as it is continuously evolving with new and exciting features added in every release.
We hope you have enjoyed this journey and look forward to hearing from you on your experience and feedback. In the Appendix, There...