[4.0] Update the docs for the scala version upgrade + missing Flink/S… #1201
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| name: "Delta Spark Master Python" | |
| on: [push, pull_request] | |
| jobs: | |
| test: | |
| name: "DSMP" | |
| runs-on: ubuntu-24.04 | |
| strategy: | |
| matrix: | |
| # These Scala versions must match those in the build.sbt | |
| scala: [2.13.13] | |
| env: | |
| SCALA_VERSION: ${{ matrix.scala }} | |
| steps: | |
| - uses: actions/checkout@v3 | |
| - uses: technote-space/get-diff-action@v4 | |
| id: git-diff | |
| with: | |
| PATTERNS: | | |
| ** | |
| .github/workflows/** | |
| !kernel/** | |
| !connectors/** | |
| - name: install java | |
| uses: actions/setup-java@v3 | |
| with: | |
| distribution: "zulu" | |
| java-version: "17" | |
| - name: Cache Scala, SBT | |
| uses: actions/cache@v3 | |
| with: | |
| path: | | |
| ~/.sbt | |
| ~/.ivy2 | |
| ~/.cache/coursier | |
| !~/.cache/coursier/v1/https/repository.apache.org/content/groups/snapshots | |
| # Change the key if dependencies are changed. For each key, GitHub Actions will cache the | |
| # the above directories when we use the key for the first time. After that, each run will | |
| # just use the cache. The cache is immutable so we need to use a new key when trying to | |
| # cache new stuff. | |
| key: delta-sbt-cache-spark-master-scala${{ matrix.scala }} | |
| - name: Install Job dependencies | |
| # TODO: update pyspark installation once Spark preview is formally released | |
| run: | | |
| sudo apt-get update | |
| sudo apt-get install -y make build-essential libssl-dev zlib1g-dev libbz2-dev libreadline-dev libsqlite3-dev wget curl llvm libncurses5-dev libncursesw5-dev xz-utils tk-dev libffi-dev liblzma-dev python3-openssl git | |
| sudo apt install libedit-dev | |
| curl -LO https://github.com/bufbuild/buf/releases/download/v1.28.1/buf-Linux-x86_64.tar.gz | |
| mkdir -p ~/buf | |
| tar -xvzf buf-Linux-x86_64.tar.gz -C ~/buf --strip-components 1 | |
| rm buf-Linux-x86_64.tar.gz | |
| sudo apt install python3-pip --fix-missing | |
| sudo pip3 install pipenv==2024.4.1 | |
| curl https://pyenv.run | bash | |
| export PATH="~/.pyenv/bin:$PATH" | |
| eval "$(pyenv init -)" | |
| eval "$(pyenv virtualenv-init -)" | |
| pyenv install 3.9 | |
| pyenv global system 3.9 | |
| pipenv --python 3.9 install | |
| # Update the pip version to 24.0. By default `pyenv.run` installs the latest pip version | |
| # available. From version 24.1, `pip` doesn't allow installing python packages | |
| # with version string containing `-`. In Delta-Spark case, the pypi package generated has | |
| # `-SNAPSHOT` in version (e.g. `3.3.0-SNAPSHOT`) as the version is picked up from | |
| # the`version.sbt` file. | |
| pipenv run pip install pip==24.0 setuptools==69.5.1 wheel==0.43.0 | |
| pipenv run pip install flake8==3.9.0 | |
| pipenv run pip install black==23.12.1 | |
| pipenv run pip install importlib_metadata==3.10.0 | |
| pipenv run pip install mypy==1.8.0 | |
| pipenv run pip install mypy-protobuf==3.3.0 | |
| pipenv run pip install cryptography==37.0.4 | |
| pipenv run pip install twine==4.0.1 | |
| pipenv run pip install wheel==0.33.4 | |
| pipenv run pip install setuptools==41.1.0 | |
| pipenv run pip install pydocstyle==3.0.0 | |
| pipenv run pip install pandas==2.2.0 | |
| pipenv run pip install pyarrow==11.0.0 | |
| pipenv run pip install pypandoc==1.3.3 | |
| pipenv run pip install numpy==1.22.4 | |
| pipenv run pip install grpcio==1.67.0 | |
| pipenv run pip install grpcio-status==1.67.0 | |
| pipenv run pip install googleapis-common-protos==1.65.0 | |
| pipenv run pip install protobuf==5.29.1 | |
| pipenv run pip install googleapis-common-protos-stubs==2.2.0 | |
| pipenv run pip install grpc-stubs==1.24.11 | |
| pipenv run pip install pyspark==4.0.0 | |
| if: steps.git-diff.outputs.diff | |
| - name: Run Python tests | |
| # when changing TEST_PARALLELISM_COUNT make sure to also change it in spark_master_test.yaml | |
| run: | | |
| TEST_PARALLELISM_COUNT=4 USE_SPARK_MASTER=true pipenv run python run-tests.py --group spark-python | |
| if: steps.git-diff.outputs.diff |