We require a working installation of Spark. This means that you would have followed the steps outlined in Chapter 1, Installing and Configuring Spark. As a reminder, to start the PySpark shell for your local Spark cluster, you can run the following command:
./bin/pyspark --master local[n]
Where n is the number of cores.