Nothing Special   »   [go: up one dir, main page]

Download Apache Spark™

  1. Choose a Spark release:

  2. Choose a package type:

  3. Download Spark:

  4. Verify this release using the and project release KEYS by following these procedures.

Note that Spark 3 is pre-built with Scala 2.12 in general and Spark 3.2+ provides additional pre-built distribution with Scala 2.13.

Spark artifacts are hosted in Maven Central. You can add a Maven dependency with the following coordinates:

groupId: org.apache.spark
artifactId: spark-core_2.12
version: 3.5.3

Installing with PyPi

PySpark is now available in pypi. To install just run pip install pyspark.

Installing with Docker

Spark docker images are available from Dockerhub under the accounts of both The Apache Software Foundation and Official Images.

Note that, these images contain non-ASF software and may be subject to different license terms. Please check their Dockerfiles to verify whether they are compatible with your deployment.

Release notes for stable releases

    Archived releases

    As new Spark releases come out for each development stream, previous ones will be archived, but they are still available at Spark release archives.

    NOTE: Previous releases of Spark may be affected by security issues. Please consult the Security page for a list of known issues that may affect the version you download before deciding to use it.