PySpark Setup
PySpark is included with Spark. Use this for local usage or as a client to connect to a cluster instead of setting up a cluster itself. For regular users, use the Invariant JupyterLab build.
Install pyspark by using pyPI in the newly created environment. This will install PySpark under the new virtual environment.
Alternatively, install PySpark from Conda
For more details about the API, refer to Apache Spark website
Last updated