diff --git a/.github/workflows/coverage-badge.yaml b/.github/workflows/coverage-badge.yaml index 2c3b40fa..2a92010b 100644 --- a/.github/workflows/coverage-badge.yaml +++ b/.github/workflows/coverage-badge.yaml @@ -9,6 +9,9 @@ on: jobs: report: + permissions: + contents: write + pull-requests: write runs-on: ubuntu-latest steps: @@ -47,7 +50,14 @@ jobs: - name: Create Pull Request if: steps.changed_files.outputs.files_changed == 'true' - uses: peter-evans/create-pull-request@v4 + uses: peter-evans/create-pull-request@v6 with: token: ${{ secrets.GITHUB_TOKEN }} title: "[Automatic] Coverage Badge Update" + commit-message: "Updated coverage.svg" + branch: create-pull-request/coverage-badge-update + delete-branch: true + body: | + This is an automated pull request to update the coverage badge. + + - Updated coverage.svg based on latest test results diff --git a/.github/workflows/e2e_tests.yaml b/.github/workflows/e2e_tests.yaml index ba59a9e1..44bf1214 100644 --- a/.github/workflows/e2e_tests.yaml +++ b/.github/workflows/e2e_tests.yaml @@ -5,13 +5,13 @@ on: pull_request: branches: - main - - 'release-*' + - "release-*" - ray-jobs-feature paths-ignore: - - 'docs/**' - - '**.adoc' - - '**.md' - - 'LICENSE' + - "docs/**" + - "**.adoc" + - "**.md" + - "LICENSE" concurrency: group: ${{ github.head_ref }}-${{ github.workflow }} @@ -33,9 +33,9 @@ jobs: - name: Checkout common repo code uses: actions/checkout@v4 with: - repository: 'project-codeflare/codeflare-common' - ref: 'main' - path: 'common' + repository: "project-codeflare/codeflare-common" + ref: "main" + path: "common" - name: Checkout CodeFlare operator repository uses: actions/checkout@v4 @@ -46,7 +46,7 @@ jobs: - name: Set Go uses: actions/setup-go@v5 with: - go-version-file: './codeflare-operator/go.mod' + go-version-file: "./codeflare-operator/go.mod" cache-dependency-path: "./codeflare-operator/go.sum" - name: Set up gotestfmt @@ -76,7 +76,7 @@ jobs: run: | cd codeflare-operator echo Setting up CodeFlare stack - make setup-e2e + make setup-e2e KUEUE_VERSION=v0.13.4 KUBERAY_VERSION=v1.4.0 echo Deploying CodeFlare operator make deploy -e IMG="${CODEFLARE_OPERATOR_IMG}" -e ENV="e2e" kubectl wait --timeout=120s --for=condition=Available=true deployment -n openshift-operators codeflare-operator-manager @@ -95,6 +95,10 @@ jobs: kubectl create clusterrolebinding sdk-user-namespace-creator --clusterrole=namespace-creator --user=sdk-user kubectl create clusterrole raycluster-creator --verb=get,list,create,delete,patch --resource=rayclusters kubectl create clusterrolebinding sdk-user-raycluster-creator --clusterrole=raycluster-creator --user=sdk-user + kubectl create clusterrole rayjob-creator --verb=get,list,create,delete,patch --resource=rayjobs + kubectl create clusterrolebinding sdk-user-rayjob-creator --clusterrole=rayjob-creator --user=sdk-user + kubectl create clusterrole rayjob-status-reader --verb=get,list,patch,update --resource=rayjobs/status + kubectl create clusterrolebinding sdk-user-rayjob-status-reader --clusterrole=rayjob-status-reader --user=sdk-user kubectl create clusterrole appwrapper-creator --verb=get,list,create,delete,patch --resource=appwrappers kubectl create clusterrolebinding sdk-user-appwrapper-creator --clusterrole=appwrapper-creator --user=sdk-user kubectl create clusterrole resourceflavor-creator --verb=get,list,create,delete --resource=resourceflavors @@ -122,7 +126,7 @@ jobs: pip install poetry poetry install --with test,docs echo "Running e2e tests..." - poetry run pytest -v -s ./tests/e2e -m 'kind and nvidia_gpu' > ${CODEFLARE_TEST_OUTPUT_DIR}/pytest_output.log 2>&1 + poetry run pytest -v -s ./tests/e2e/ -m 'kind and nvidia_gpu' > ${CODEFLARE_TEST_OUTPUT_DIR}/pytest_output.log 2>&1 env: GRPC_DNS_RESOLVER: "native" @@ -146,7 +150,13 @@ jobs: if: always() && steps.deploy.outcome == 'success' run: | echo "Printing KubeRay operator logs" - kubectl logs -n ray-system --tail -1 -l app.kubernetes.io/name=kuberay | tee ${CODEFLARE_TEST_OUTPUT_DIR}/kuberay.log + kubectl logs -n default --tail -1 -l app.kubernetes.io/name=kuberay | tee ${CODEFLARE_TEST_OUTPUT_DIR}/kuberay.log + + - name: Print Kueue controller logs + if: always() && steps.deploy.outcome == 'success' + run: | + echo "Printing Kueue controller logs" + kubectl logs -n kueue-system --tail -1 -l control-plane=controller-manager | tee ${CODEFLARE_TEST_OUTPUT_DIR}/kueue.log - name: Export all KinD pod logs uses: ./common/github-actions/kind-export-logs diff --git a/.github/workflows/rayjob_e2e_tests.yaml b/.github/workflows/rayjob_e2e_tests.yaml new file mode 100644 index 00000000..ba0659c0 --- /dev/null +++ b/.github/workflows/rayjob_e2e_tests.yaml @@ -0,0 +1,176 @@ +# rayjob e2e tests workflow for CodeFlare-SDK +name: rayjob-e2e + +on: + pull_request: + branches: + - main + - "release-*" + - ray-jobs-feature + paths-ignore: + - "docs/**" + - "**.adoc" + - "**.md" + - "LICENSE" + +concurrency: + group: ${{ github.head_ref }}-${{ github.workflow }} + cancel-in-progress: true + +env: + CODEFLARE_OPERATOR_IMG: "quay.io/project-codeflare/codeflare-operator:dev" + +jobs: + kubernetes-rayjob: + runs-on: gpu-t4-4-core + + steps: + - name: Checkout code + uses: actions/checkout@v4 + with: + submodules: recursive + + - name: Checkout common repo code + uses: actions/checkout@v4 + with: + repository: "project-codeflare/codeflare-common" + ref: "main" + path: "common" + + - name: Checkout CodeFlare operator repository + uses: actions/checkout@v4 + with: + repository: project-codeflare/codeflare-operator + path: codeflare-operator + + - name: Set Go + uses: actions/setup-go@v5 + with: + go-version-file: "./codeflare-operator/go.mod" + cache-dependency-path: "./codeflare-operator/go.sum" + + - name: Set up gotestfmt + uses: gotesttools/gotestfmt-action@v2 + with: + token: ${{ secrets.GITHUB_TOKEN }} + + - name: Set up specific Python version + uses: actions/setup-python@v5 + with: + python-version: "3.11" + cache: "pip" # caching pip dependencies + + - name: Setup NVidia GPU environment for KinD + uses: ./common/github-actions/nvidia-gpu-setup + + - name: Setup and start KinD cluster + uses: ./common/github-actions/kind + with: + worker-nodes: 1 + + - name: Install NVidia GPU operator for KinD + uses: ./common/github-actions/nvidia-gpu-operator + + - name: Deploy CodeFlare stack + id: deploy + run: | + cd codeflare-operator + echo Setting up CodeFlare stack + make setup-e2e KUEUE_VERSION=v0.13.4 KUBERAY_VERSION=v1.4.0 + echo Deploying CodeFlare operator + make deploy -e IMG="${CODEFLARE_OPERATOR_IMG}" -e ENV="e2e" + kubectl wait --timeout=120s --for=condition=Available=true deployment -n openshift-operators codeflare-operator-manager + cd .. + + - name: Add user to KinD + uses: ./common/github-actions/kind-add-user + with: + user-name: sdk-user + + - name: Configure RBAC for sdk user with limited permissions + run: | + kubectl create clusterrole list-ingresses --verb=get,list --resource=ingresses + kubectl create clusterrolebinding sdk-user-list-ingresses --clusterrole=list-ingresses --user=sdk-user + kubectl create clusterrole namespace-creator --verb=get,list,create,delete,patch --resource=namespaces + kubectl create clusterrolebinding sdk-user-namespace-creator --clusterrole=namespace-creator --user=sdk-user + kubectl create clusterrole raycluster-creator --verb=get,list,create,delete,patch --resource=rayclusters + kubectl create clusterrolebinding sdk-user-raycluster-creator --clusterrole=raycluster-creator --user=sdk-user + kubectl create clusterrole rayjob-creator --verb=get,list,create,delete,patch --resource=rayjobs + kubectl create clusterrolebinding sdk-user-rayjob-creator --clusterrole=rayjob-creator --user=sdk-user + kubectl create clusterrole rayjob-status-reader --verb=get,list,patch,update --resource=rayjobs/status + kubectl create clusterrolebinding sdk-user-rayjob-status-reader --clusterrole=rayjob-status-reader --user=sdk-user + kubectl create clusterrole appwrapper-creator --verb=get,list,create,delete,patch --resource=appwrappers + kubectl create clusterrolebinding sdk-user-appwrapper-creator --clusterrole=appwrapper-creator --user=sdk-user + kubectl create clusterrole resourceflavor-creator --verb=get,list,create,delete --resource=resourceflavors + kubectl create clusterrolebinding sdk-user-resourceflavor-creator --clusterrole=resourceflavor-creator --user=sdk-user + kubectl create clusterrole clusterqueue-creator --verb=get,list,create,delete,patch --resource=clusterqueues + kubectl create clusterrolebinding sdk-user-clusterqueue-creator --clusterrole=clusterqueue-creator --user=sdk-user + kubectl create clusterrole localqueue-creator --verb=get,list,create,delete,patch --resource=localqueues + kubectl create clusterrolebinding sdk-user-localqueue-creator --clusterrole=localqueue-creator --user=sdk-user + kubectl create clusterrole list-secrets --verb=get,list --resource=secrets + kubectl create clusterrolebinding sdk-user-list-secrets --clusterrole=list-secrets --user=sdk-user + kubectl create clusterrole pod-creator --verb=get,list,watch --resource=pods + kubectl create clusterrolebinding sdk-user-pod-creator --clusterrole=pod-creator --user=sdk-user + kubectl create clusterrole service-reader --verb=get,list,watch --resource=services + kubectl create clusterrolebinding sdk-user-service-reader --clusterrole=service-reader --user=sdk-user + kubectl create clusterrole port-forward-pods --verb=create --resource=pods/portforward + kubectl create clusterrolebinding sdk-user-port-forward-pods-binding --clusterrole=port-forward-pods --user=sdk-user + kubectl create clusterrole secret-manager --verb=get,list,create,delete,update,patch --resource=secrets + kubectl create clusterrolebinding sdk-user-secret-manager --clusterrole=secret-manager --user=sdk-user + kubectl create clusterrole workload-reader --verb=get,list,watch --resource=workloads + kubectl create clusterrolebinding sdk-user-workload-reader --clusterrole=workload-reader --user=sdk-user + kubectl config use-context sdk-user + + - name: Run RayJob E2E tests + run: | + export CODEFLARE_TEST_OUTPUT_DIR=${{ env.TEMP_DIR }} + echo "CODEFLARE_TEST_OUTPUT_DIR=${CODEFLARE_TEST_OUTPUT_DIR}" >> $GITHUB_ENV + + set -euo pipefail + pip install poetry + poetry install --with test,docs + echo "Running RayJob e2e tests..." + poetry run pytest -v -s ./tests/e2e/rayjob/ > ${CODEFLARE_TEST_OUTPUT_DIR}/pytest_output_rayjob.log 2>&1 + + - name: Switch to kind-cluster context to print logs + if: always() && steps.deploy.outcome == 'success' + run: kubectl config use-context kind-cluster + + - name: Print Pytest output log + if: always() && steps.deploy.outcome == 'success' + run: | + echo "Printing Pytest output logs" + cat ${CODEFLARE_TEST_OUTPUT_DIR}/pytest_output_rayjob.log + + - name: Print CodeFlare operator logs + if: always() && steps.deploy.outcome == 'success' + run: | + echo "Printing CodeFlare operator logs" + kubectl logs -n openshift-operators --tail -1 -l app.kubernetes.io/name=codeflare-operator | tee ${CODEFLARE_TEST_OUTPUT_DIR}/codeflare-operator.log + + - name: Print KubeRay operator logs + if: always() && steps.deploy.outcome == 'success' + run: | + echo "Printing KubeRay operator logs" + kubectl logs -n default --tail -1 -l app.kubernetes.io/name=kuberay | tee ${CODEFLARE_TEST_OUTPUT_DIR}/kuberay.log + + - name: Print Kueue controller logs + if: always() && steps.deploy.outcome == 'success' + run: | + echo "Printing Kueue controller logs" + kubectl logs -n kueue-system --tail -1 -l control-plane=controller-manager | tee ${CODEFLARE_TEST_OUTPUT_DIR}/kueue.log + + - name: Export all KinD pod logs + uses: ./common/github-actions/kind-export-logs + if: always() && steps.deploy.outcome == 'success' + with: + output-directory: ${CODEFLARE_TEST_OUTPUT_DIR} + + - name: Upload logs + uses: actions/upload-artifact@v4 + if: always() && steps.deploy.outcome == 'success' + with: + name: logs + retention-days: 10 + path: | + ${{ env.CODEFLARE_TEST_OUTPUT_DIR }}/**/*.log diff --git a/codecov.yml b/codecov.yml index 550965e6..4494dcd4 100644 --- a/codecov.yml +++ b/codecov.yml @@ -1,3 +1,17 @@ ignore: - "**/*.ipynb" - "demo-notebooks/**" + - "**/__init__.py" + +coverage: + precision: 2 + round: down + status: + project: + default: + target: auto + threshold: 2.5% + patch: + default: + target: 85% + threshold: 2.5% diff --git a/coverage.svg b/coverage.svg index a8c7e72a..59d64b37 100644 --- a/coverage.svg +++ b/coverage.svg @@ -15,7 +15,7 @@ coverage coverage - 92% - 92% + 93% + 93% diff --git a/demo-notebooks/guided-demos/4_rayjob_existing_cluster.ipynb b/demo-notebooks/guided-demos/4_rayjob_existing_cluster.ipynb new file mode 100644 index 00000000..c0737db0 --- /dev/null +++ b/demo-notebooks/guided-demos/4_rayjob_existing_cluster.ipynb @@ -0,0 +1,207 @@ +{ + "cells": [ + { + "cell_type": "markdown", + "id": "9259e514", + "metadata": {}, + "source": [ + "# Submitting RayJobs against an existing RayCluster\n", + "\n", + "In this notebook, we will go through the basics of using the SDK to:\n", + " * Spin up a Ray cluster with our desired resources\n", + " * Verify the status of this cluster\n", + " * Submit a RayJob against that cluster\n", + " * Verify the status of this job" + ] + }, + { + "cell_type": "markdown", + "id": "18136ea7", + "metadata": {}, + "source": [ + "## Creating the RayCluster" + ] + }, + { + "cell_type": "markdown", + "id": "a1c2545d", + "metadata": {}, + "source": [ + "First, we'll need to import the relevant CodeFlare SDK packages. You can do this by executing the below cell." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "51e18292", + "metadata": {}, + "outputs": [], + "source": [ + "from codeflare_sdk import Cluster, ClusterConfiguration, RayJob" + ] + }, + { + "cell_type": "markdown", + "id": "649c5911", + "metadata": {}, + "source": [ + "Run the below `oc login` command using your Token and Server URL. Ensure the command is prepended by `!` and not `%`. This will work when running both locally and within RHOAI." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "dc364888", + "metadata": {}, + "outputs": [], + "source": [ + "!oc login --token= --server=" + ] + }, + { + "cell_type": "markdown", + "id": "5581eca9", + "metadata": {}, + "source": [ + "Next we'll need to initalize our RayCluster and apply it. You can do this be executing the below cell." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "3094c60a", + "metadata": {}, + "outputs": [], + "source": [ + "cluster = Cluster(ClusterConfiguration(\n", + " name='rayjob-cluster',\n", + " head_extended_resource_requests={'nvidia.com/gpu':0},\n", + " worker_extended_resource_requests={'nvidia.com/gpu':0},\n", + " num_workers=2,\n", + " worker_cpu_requests=1,\n", + " worker_cpu_limits=1,\n", + " worker_memory_requests=4,\n", + " worker_memory_limits=4,\n", + "\n", + "))\n", + "\n", + "cluster.apply()" + ] + }, + { + "cell_type": "markdown", + "id": "f3612de2", + "metadata": {}, + "source": [ + "We can check the status of our cluster by executing the below cell. If it's not up immediately, run the cell a few more times until you see that it's in a 'running' state." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "96d92f93", + "metadata": {}, + "outputs": [], + "source": [ + "cluster.status()" + ] + }, + { + "cell_type": "markdown", + "id": "a0e2a650", + "metadata": {}, + "source": [ + "## Submitting the RayJob" + ] + }, + { + "cell_type": "markdown", + "id": "4cf03419", + "metadata": {}, + "source": [ + "Now we can create the RayJob that we want to submit against the running cluster. The process is quite similar to how we initialize and apply the cluster. \n", + "In this context, we need to use the `cluster_name` variable to point it to our existing cluster.\n", + "\n", + "For the sake of demonstration, the job we'll submit via the `entrypoint` is a single python command. In standard practice this would be pointed to a python training script.\n", + "\n", + "We'll then call the `submit()` function to run the job against our cluster.\n", + "\n", + "You can run the below cell to achieve this." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "94edca70", + "metadata": {}, + "outputs": [], + "source": [ + "rayjob = RayJob(\n", + " job_name=\"sdk-test-job\",\n", + " cluster_name=\"rayjob-cluster\",\n", + " namespace=\"your-namespace\",\n", + " entrypoint=\"python -c 'import time; time.sleep(20)'\",\n", + ")\n", + "\n", + "rayjob.submit()" + ] + }, + { + "cell_type": "markdown", + "id": "30a8899a", + "metadata": {}, + "source": [ + "We can observe the status of the RayJob in the same way as the RayCluster by invoking the `status()` function via the below cell." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "3283b09c", + "metadata": {}, + "outputs": [], + "source": [ + "rayjob.status()" + ] + }, + { + "cell_type": "markdown", + "id": "9f3c9c9f", + "metadata": {}, + "source": [ + "This function will output different tables based on the RayJob's current status. You can re-run the cell multiple times to observe the changes as you need to. Once you've observed that the job has been completed, you can shut down the cluster we created earlier by executing the below cell." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "5b11e379", + "metadata": {}, + "outputs": [], + "source": [ + "cluster.down()" + ] + } + ], + "metadata": { + "kernelspec": { + "display_name": "Python 3", + "language": "python", + "name": "python3" + }, + "language_info": { + "codemirror_mode": { + "name": "ipython", + "version": 3 + }, + "file_extension": ".py", + "mimetype": "text/x-python", + "name": "python", + "nbconvert_exporter": "python", + "pygments_lexer": "ipython3", + "version": "3.11.11" + } + }, + "nbformat": 4, + "nbformat_minor": 5 +} diff --git a/demo-notebooks/guided-demos/5_submit_rayjob_cr.ipynb b/demo-notebooks/guided-demos/5_submit_rayjob_cr.ipynb new file mode 100644 index 00000000..1d9630b7 --- /dev/null +++ b/demo-notebooks/guided-demos/5_submit_rayjob_cr.ipynb @@ -0,0 +1,144 @@ +{ + "cells": [ + { + "cell_type": "markdown", + "id": "9259e514", + "metadata": {}, + "source": [ + "# Submitting a RayJob CR\n", + "\n", + "In this notebook, we will go through the basics of using the SDK to:\n", + " * Define a RayCluster configuration\n", + " * Use this configuration alongside a RayJob definition\n", + " * Submit the RayJob, and allow Kuberay Operator to lifecycle the RayCluster for the RayJob" + ] + }, + { + "cell_type": "markdown", + "id": "18136ea7", + "metadata": {}, + "source": [ + "## Defining and Submitting the RayJob\n", + "First, we'll need to import the relevant CodeFlare SDK packages. You can do this by executing the below cell." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "51e18292", + "metadata": {}, + "outputs": [], + "source": [ + "from codeflare_sdk import RayJob, ManagedClusterConfig" + ] + }, + { + "cell_type": "markdown", + "id": "649c5911", + "metadata": {}, + "source": [ + "Run the below `oc login` command using your Token and Server URL. Ensure the command is prepended by `!` and not `%`. This will work when running both locally and within RHOAI." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "dc364888", + "metadata": {}, + "outputs": [], + "source": [ + "!oc login --token= --server=" + ] + }, + { + "cell_type": "markdown", + "id": "5581eca9", + "metadata": {}, + "source": [ + "Next we'll need to define the ManagedClusterConfig. Kuberay will use this to spin up a short-lived RayCluster that will only exist as long as the job" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "3094c60a", + "metadata": {}, + "outputs": [], + "source": [ + "cluster_config = ManagedClusterConfig(\n", + " num_workers=2,\n", + " worker_cpu_requests=1,\n", + " worker_cpu_limits=1,\n", + " worker_memory_requests=4,\n", + " worker_memory_limits=4,\n", + " head_accelerators={'nvidia.com/gpu': 0},\n", + " worker_accelerators={'nvidia.com/gpu': 0},\n", + ")" + ] + }, + { + "cell_type": "markdown", + "id": "02a2b32b", + "metadata": {}, + "source": [ + "Lastly we can pass the ManagedClusterConfig into the RayJob and submit it. You do not need to worry about tearing down the cluster when the job has completed, that is handled for you!" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "e905ccea", + "metadata": {}, + "outputs": [], + "source": [ + "job = RayJob(\n", + " job_name=\"demo-rayjob\",\n", + " entrypoint=\"python -c 'print(\\\"Hello from RayJob!\\\")'\",\n", + " cluster_config=cluster_config,\n", + " namespace=\"your-namespace\"\n", + ")\n", + "\n", + "job.submit()" + ] + }, + { + "cell_type": "markdown", + "id": "f3612de2", + "metadata": {}, + "source": [ + "We can check the status of our job by executing the below cell. The status may appear as `unknown` for a time while the RayCluster spins up." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "96d92f93", + "metadata": {}, + "outputs": [], + "source": [ + "job.status()" + ] + } + ], + "metadata": { + "kernelspec": { + "display_name": "Python 3", + "language": "python", + "name": "python3" + }, + "language_info": { + "codemirror_mode": { + "name": "ipython", + "version": 3 + }, + "file_extension": ".py", + "mimetype": "text/x-python", + "name": "python", + "nbconvert_exporter": "python", + "pygments_lexer": "ipython3", + "version": "3.11.11" + } + }, + "nbformat": 4, + "nbformat_minor": 5 +} diff --git a/demo-notebooks/guided-demos/preview_nbs/4_rayjob_existing_cluster.ipynb b/demo-notebooks/guided-demos/preview_nbs/4_rayjob_existing_cluster.ipynb new file mode 100644 index 00000000..5348099c --- /dev/null +++ b/demo-notebooks/guided-demos/preview_nbs/4_rayjob_existing_cluster.ipynb @@ -0,0 +1,212 @@ +{ + "cells": [ + { + "cell_type": "markdown", + "id": "9259e514", + "metadata": {}, + "source": [ + "# Submitting RayJobs against an existing RayCluster\n", + "\n", + "In this notebook, we will go through the basics of using the SDK to:\n", + " * Spin up a Ray cluster with our desired resources\n", + " * Verify the status of this cluster\n", + " * Submit a RayJob against that cluster\n", + " * Verify the status of this job" + ] + }, + { + "cell_type": "markdown", + "id": "18136ea7", + "metadata": {}, + "source": [ + "## Creating the RayCluster" + ] + }, + { + "cell_type": "markdown", + "id": "a1c2545d", + "metadata": {}, + "source": [ + "First, we'll need to import the relevant CodeFlare SDK packages. You can do this by executing the below cell." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "51e18292", + "metadata": {}, + "outputs": [], + "source": [ + "from codeflare_sdk import Cluster, ClusterConfiguration, RayJob, TokenAuthentication" + ] + }, + { + "cell_type": "markdown", + "id": "649c5911", + "metadata": {}, + "source": [ + "Execute the below cell to authenticate the notebook via OpenShift." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "dc364888", + "metadata": {}, + "outputs": [], + "source": [ + "auth = TokenAuthentication(\n", + " token = \"XXXXX\",\n", + " server = \"XXXXX\",\n", + " skip_tls=False\n", + ")\n", + "auth.login()" + ] + }, + { + "cell_type": "markdown", + "id": "5581eca9", + "metadata": {}, + "source": [ + "Next we'll need to initalize our RayCluster and apply it. You can do this be executing the below cell." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "3094c60a", + "metadata": {}, + "outputs": [], + "source": [ + "cluster = Cluster(ClusterConfiguration(\n", + " name='rayjob-cluster',\n", + " head_extended_resource_requests={'nvidia.com/gpu':0},\n", + " worker_extended_resource_requests={'nvidia.com/gpu':0},\n", + " num_workers=2,\n", + " worker_cpu_requests=1,\n", + " worker_cpu_limits=1,\n", + " worker_memory_requests=4,\n", + " worker_memory_limits=4,\n", + "\n", + "))\n", + "\n", + "cluster.apply()" + ] + }, + { + "cell_type": "markdown", + "id": "f3612de2", + "metadata": {}, + "source": [ + "We can check the status of our cluster by executing the below cell. If it's not up immediately, run the cell a few more times until you see that it's in a 'running' state." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "96d92f93", + "metadata": {}, + "outputs": [], + "source": [ + "cluster.status()" + ] + }, + { + "cell_type": "markdown", + "id": "a0e2a650", + "metadata": {}, + "source": [ + "## Creating and Submitting the RayJob" + ] + }, + { + "cell_type": "markdown", + "id": "4cf03419", + "metadata": {}, + "source": [ + "Now we can create the RayJob that we want to submit against the running cluster. The process is quite similar to how we initialize and apply the cluster. \n", + "In this context, we need to use the `cluster_name` variable to point it to our existing cluster.\n", + "\n", + "For the sake of demonstration, the job we'll submit via the `entrypoint` is a single python command. In standard practice this would be pointed to a python training script.\n", + "\n", + "We'll then call the `submit()` function to run the job against our cluster.\n", + "\n", + "You can run the below cell to achieve this." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "94edca70", + "metadata": {}, + "outputs": [], + "source": [ + "rayjob = RayJob(\n", + " job_name=\"sdk-test-job\",\n", + " cluster_name=\"rayjob-cluster\",\n", + " namespace=\"rhods-notebooks\",\n", + " entrypoint=\"python -c 'import time; time.sleep(20)'\",\n", + ")\n", + "\n", + "rayjob.submit()" + ] + }, + { + "cell_type": "markdown", + "id": "30a8899a", + "metadata": {}, + "source": [ + "We can observe the status of the RayJob in the same way as the RayCluster by invoking the `submit()` function via the below cell." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "3283b09c", + "metadata": {}, + "outputs": [], + "source": [ + "rayjob.submit()" + ] + }, + { + "cell_type": "markdown", + "id": "9f3c9c9f", + "metadata": {}, + "source": [ + "This function will output different tables based on the RayJob's current status. You can re-run the cell multiple times to observe the changes as you need to. Once you've observed that the job has been completed, you can shut down the cluster we created earlier by executing the below cell." + ] + }, + { + "cell_type": "code", + "execution_count": null, + "id": "5b11e379", + "metadata": {}, + "outputs": [], + "source": [ + "cluster.down()" + ] + } + ], + "metadata": { + "kernelspec": { + "display_name": "Python 3", + "language": "python", + "name": "python3" + }, + "language_info": { + "codemirror_mode": { + "name": "ipython", + "version": 3 + }, + "file_extension": ".py", + "mimetype": "text/x-python", + "name": "python", + "nbconvert_exporter": "python", + "pygments_lexer": "ipython3", + "version": "3.11.11" + } + }, + "nbformat": 4, + "nbformat_minor": 5 +} diff --git a/poetry.lock b/poetry.lock index 49e45352..88ceb3cb 100644 --- a/poetry.lock +++ b/poetry.lock @@ -1,4 +1,4 @@ -# This file is automatically @generated by Poetry 2.1.1 and should not be changed by hand. +# This file is automatically @generated by Poetry 2.1.4 and should not be changed by hand. [[package]] name = "aiohappyeyeballs" @@ -14,103 +14,103 @@ files = [ [[package]] name = "aiohttp" -version = "3.12.12" +version = "3.12.14" description = "Async http client/server framework (asyncio)" optional = false python-versions = ">=3.9" groups = ["main"] files = [ - {file = "aiohttp-3.12.12-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:6f25e9d274d6abbb15254f76f100c3984d6b9ad6e66263cc60a465dd5c7e48f5"}, - {file = "aiohttp-3.12.12-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b8ec3c1a1c13d24941b5b913607e57b9364e4c0ea69d5363181467492c4b2ba6"}, - {file = "aiohttp-3.12.12-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:81ef2f9253c327c211cb7b06ea2edd90e637cf21c347b894d540466b8d304e08"}, - {file = "aiohttp-3.12.12-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:28ded835c3663fd41c9ad44685811b11e34e6ac9a7516a30bfce13f6abba4496"}, - {file = "aiohttp-3.12.12-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:a4b78ccf254fc10605b263996949a94ca3f50e4f9100e05137d6583e266b711e"}, - {file = "aiohttp-3.12.12-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4f4a5af90d5232c41bb857568fe7d11ed84408653ec9da1ff999cc30258b9bd1"}, - {file = "aiohttp-3.12.12-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:ffa5205c2f53f1120e93fdf2eca41b0f6344db131bc421246ee82c1e1038a14a"}, - {file = "aiohttp-3.12.12-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f68301660f0d7a3eddfb84f959f78a8f9db98c76a49b5235508fa16edaad0f7c"}, - {file = "aiohttp-3.12.12-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:db874d3b0c92fdbb553751af9d2733b378c25cc83cd9dfba87f12fafd2dc9cd5"}, - {file = "aiohttp-3.12.12-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:5e53cf9c201b45838a2d07b1f2d5f7fec9666db7979240002ce64f9b8a1e0cf2"}, - {file = "aiohttp-3.12.12-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:8687cc5f32b4e328c233acd387d09a1b477007896b2f03c1c823a0fd05f63883"}, - {file = "aiohttp-3.12.12-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:5ee537ad29de716a3d8dc46c609908de0c25ffeebf93cd94a03d64cdc07d66d0"}, - {file = "aiohttp-3.12.12-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:411f821be5af6af11dc5bed6c6c1dc6b6b25b91737d968ec2756f9baa75e5f9b"}, - {file = "aiohttp-3.12.12-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:f90319d94cf5f9786773237f24bd235a7b5959089f1af8ec1154580a3434b503"}, - {file = "aiohttp-3.12.12-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:73b148e606f34e9d513c451fd65efe1091772659ca5703338a396a99f60108ff"}, - {file = "aiohttp-3.12.12-cp310-cp310-win32.whl", hash = "sha256:d40e7bfd577fdc8a92b72f35dfbdd3ec90f1bc8a72a42037fefe34d4eca2d4a1"}, - {file = "aiohttp-3.12.12-cp310-cp310-win_amd64.whl", hash = "sha256:65c7804a2343893d6dea9fce69811aea0a9ac47f68312cf2e3ee1668cd9a387f"}, - {file = "aiohttp-3.12.12-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:38823fe0d8bc059b3eaedb263fe427d887c7032e72b4ef92c472953285f0e658"}, - {file = "aiohttp-3.12.12-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:10237f2c34711215d04ed21da63852ce023608299554080a45c576215d9df81c"}, - {file = "aiohttp-3.12.12-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:563ec477c0dc6d56fc7f943a3475b5acdb399c7686c30f5a98ada24bb7562c7a"}, - {file = "aiohttp-3.12.12-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f3d05c46a61aca7c47df74afff818bc06a251ab95d95ff80b53665edfe1e0bdf"}, - {file = "aiohttp-3.12.12-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:277c882916759b4a6b6dc7e2ceb124aad071b3c6456487808d9ab13e1b448d57"}, - {file = "aiohttp-3.12.12-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:216abf74b324b0f4e67041dd4fb2819613909a825904f8a51701fbcd40c09cd7"}, - {file = "aiohttp-3.12.12-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:65d6cefad286459b68e7f867b9586a821fb7f121057b88f02f536ef570992329"}, - {file = "aiohttp-3.12.12-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:feaaaff61966b5f4b4eae0b79fc79427f49484e4cfa5ab7d138ecd933ab540a8"}, - {file = "aiohttp-3.12.12-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a05917780b7cad1755784b16cfaad806bc16029a93d15f063ca60185b7d9ba05"}, - {file = "aiohttp-3.12.12-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:082c5ec6d262c1b2ee01c63f4fb9152c17f11692bf16f0f100ad94a7a287d456"}, - {file = "aiohttp-3.12.12-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:b265a3a8b379b38696ac78bdef943bdc4f4a5d6bed1a3fb5c75c6bab1ecea422"}, - {file = "aiohttp-3.12.12-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:2e0f2e208914ecbc4b2a3b7b4daa759d0c587d9a0b451bb0835ac47fae7fa735"}, - {file = "aiohttp-3.12.12-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:9923b025845b72f64d167bca221113377c8ffabd0a351dc18fb839d401ee8e22"}, - {file = "aiohttp-3.12.12-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:1ebb213445900527831fecc70e185bf142fdfe5f2a691075f22d63c65ee3c35a"}, - {file = "aiohttp-3.12.12-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:6fc369fb273a8328077d37798b77c1e65676709af5c182cb74bd169ca9defe81"}, - {file = "aiohttp-3.12.12-cp311-cp311-win32.whl", hash = "sha256:58ecd10fda6a44c311cd3742cfd2aea8c4c600338e9f27cb37434d9f5ca9ddaa"}, - {file = "aiohttp-3.12.12-cp311-cp311-win_amd64.whl", hash = "sha256:b0066e88f30be00badffb5ef8f2281532b9a9020863d873ae15f7c147770b6ec"}, - {file = "aiohttp-3.12.12-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:98451ce9ce229d092f278a74a7c2a06b3aa72984673c87796126d7ccade893e9"}, - {file = "aiohttp-3.12.12-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:adbac7286d89245e1aff42e948503fdc6edf6d5d65c8e305a67c40f6a8fb95f4"}, - {file = "aiohttp-3.12.12-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:0728882115bfa85cbd8d0f664c8ccc0cfd5bd3789dd837596785450ae52fac31"}, - {file = "aiohttp-3.12.12-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6bf3b9d9e767f9d0e09fb1a31516410fc741a62cc08754578c40abc497d09540"}, - {file = "aiohttp-3.12.12-cp312-cp312-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:c944860e86b9f77a462321a440ccf6fa10f5719bb9d026f6b0b11307b1c96c7b"}, - {file = "aiohttp-3.12.12-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3b1979e1f0c98c06fd0cd940988833b102fa3aa56751f6c40ffe85cabc51f6fd"}, - {file = "aiohttp-3.12.12-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:120b7dd084e96cfdad85acea2ce1e7708c70a26db913eabb8d7b417c728f5d84"}, - {file = "aiohttp-3.12.12-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0e58f5ae79649ffa247081c2e8c85e31d29623cf2a3137dda985ae05c9478aae"}, - {file = "aiohttp-3.12.12-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:9aa5f049e3e2745b0141f13e5a64e7c48b1a1427ed18bbb7957b348f282fee56"}, - {file = "aiohttp-3.12.12-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:7163cc9cf3722d90f1822f8a38b211e3ae2fc651c63bb55449f03dc1b3ff1d44"}, - {file = "aiohttp-3.12.12-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:ef97c4d035b721de6607f3980fa3e4ef0ec3aca76474b5789b7fac286a8c4e23"}, - {file = "aiohttp-3.12.12-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:1c14448d6a86acadc3f7b2f4cc385d1fb390acb6f37dce27f86fe629410d92e3"}, - {file = "aiohttp-3.12.12-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:a1b6df6255cfc493454c79221183d64007dd5080bcda100db29b7ff181b8832c"}, - {file = "aiohttp-3.12.12-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:60fc7338dfb0626c2927bfbac4785de3ea2e2bbe3d328ba5f3ece123edda4977"}, - {file = "aiohttp-3.12.12-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:d2afc72207ef4c9d4ca9fcd00689a6a37ef2d625600c3d757b5c2b80c9d0cf9a"}, - {file = "aiohttp-3.12.12-cp312-cp312-win32.whl", hash = "sha256:8098a48f93b2cbcdb5778e7c9a0e0375363e40ad692348e6e65c3b70d593b27c"}, - {file = "aiohttp-3.12.12-cp312-cp312-win_amd64.whl", hash = "sha256:d1c1879b2e0fc337d7a1b63fe950553c2b9e93c071cf95928aeea1902d441403"}, - {file = "aiohttp-3.12.12-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:ea5d604318234427929d486954e3199aded65f41593ac57aa0241ab93dda3d15"}, - {file = "aiohttp-3.12.12-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:e03ff38250b8b572dce6fcd7b6fb6ee398bb8a59e6aa199009c5322d721df4fc"}, - {file = "aiohttp-3.12.12-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:71125b1fc2b6a94bccc63bbece620906a4dead336d2051f8af9cbf04480bc5af"}, - {file = "aiohttp-3.12.12-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:784a66f9f853a22c6b8c2bd0ff157f9b879700f468d6d72cfa99167df08c5c9c"}, - {file = "aiohttp-3.12.12-cp313-cp313-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:a5be0b58670b54301404bd1840e4902570a1c3be00358e2700919cb1ea73c438"}, - {file = "aiohttp-3.12.12-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ce8f13566fc7bf5a728275b434bc3bdea87a7ed3ad5f734102b02ca59d9b510f"}, - {file = "aiohttp-3.12.12-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d736e57d1901683bc9be648aa308cb73e646252c74b4c639c35dcd401ed385ea"}, - {file = "aiohttp-3.12.12-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e2007eaa7aae9102f211c519d1ec196bd3cecb1944a095db19eeaf132b798738"}, - {file = "aiohttp-3.12.12-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2a813e61583cab6d5cdbaa34bc28863acdb92f9f46e11de1b3b9251a1e8238f6"}, - {file = "aiohttp-3.12.12-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:e408293aa910b0aea48b86a28eace41d497a85ba16c20f619f0c604597ef996c"}, - {file = "aiohttp-3.12.12-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:f3d31faf290f5a30acba46b388465b67c6dbe8655d183e9efe2f6a1d594e6d9d"}, - {file = "aiohttp-3.12.12-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:0b84731697325b023902aa643bd1726d999f5bc7854bc28b17ff410a81151d4b"}, - {file = "aiohttp-3.12.12-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:a324c6852b6e327811748446e56cc9bb6eaa58710557922183175816e82a4234"}, - {file = "aiohttp-3.12.12-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:22fd867fbd72612dcf670c90486dbcbaf702cb807fb0b42bc0b7a142a573574a"}, - {file = "aiohttp-3.12.12-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:3e092f1a970223794a4bf620a26c0e4e4e8e36bccae9b0b5da35e6d8ee598a03"}, - {file = "aiohttp-3.12.12-cp313-cp313-win32.whl", hash = "sha256:7f5f5eb8717ef8ba15ab35fcde5a70ad28bbdc34157595d1cddd888a985f5aae"}, - {file = "aiohttp-3.12.12-cp313-cp313-win_amd64.whl", hash = "sha256:ace2499bdd03c329c054dc4b47361f2b19d5aa470f7db5c7e0e989336761b33c"}, - {file = "aiohttp-3.12.12-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:0d0b1c27c05a7d39a50e946ec5f94c3af4ffadd33fa5f20705df42fb0a72ca14"}, - {file = "aiohttp-3.12.12-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:e5928847e6f7b7434921fbabf73fa5609d1f2bf4c25d9d4522b1fcc3b51995cb"}, - {file = "aiohttp-3.12.12-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:7678147c3c85a7ae61559b06411346272ed40a08f54bc05357079a63127c9718"}, - {file = "aiohttp-3.12.12-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f50057f36f2a1d8e750b273bb966bec9f69ee1e0a20725ae081610501f25d555"}, - {file = "aiohttp-3.12.12-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:e5e834f0f11ff5805d11f0f22b627c75eadfaf91377b457875e4e3affd0b924f"}, - {file = "aiohttp-3.12.12-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:f94b2e2dea19d09745ef02ed483192260750f18731876a5c76f1c254b841443a"}, - {file = "aiohttp-3.12.12-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b434bfb49564dc1c318989a0ab1d3000d23e5cfd00d8295dc9d5a44324cdd42d"}, - {file = "aiohttp-3.12.12-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8ed76bc80177ddb7c5c93e1a6440b115ed2c92a3063420ac55206fd0832a6459"}, - {file = "aiohttp-3.12.12-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e1282a9acd378f2aed8dc79c01e702b1d5fd260ad083926a88ec7e987c4e0ade"}, - {file = "aiohttp-3.12.12-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:09a213c13fba321586edab1528b530799645b82bd64d79b779eb8d47ceea155a"}, - {file = "aiohttp-3.12.12-cp39-cp39-musllinux_1_2_armv7l.whl", hash = "sha256:72eae16a9233561d315e72ae78ed9fc65ab3db0196e56cb2d329c755d694f137"}, - {file = "aiohttp-3.12.12-cp39-cp39-musllinux_1_2_i686.whl", hash = "sha256:f25990c507dbbeefd5a6a17df32a4ace634f7b20a38211d1b9609410c7f67a24"}, - {file = "aiohttp-3.12.12-cp39-cp39-musllinux_1_2_ppc64le.whl", hash = "sha256:3a2aa255417c8ccf1b39359cd0a3d63ae3b5ced83958dbebc4d9113327c0536a"}, - {file = "aiohttp-3.12.12-cp39-cp39-musllinux_1_2_s390x.whl", hash = "sha256:a4c53b89b3f838e9c25f943d1257efff10b348cb56895f408ddbcb0ec953a2ad"}, - {file = "aiohttp-3.12.12-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:b5a49c2dcb32114455ad503e8354624d85ab311cbe032da03965882492a9cb98"}, - {file = "aiohttp-3.12.12-cp39-cp39-win32.whl", hash = "sha256:74fddc0ba8cea6b9c5bd732eb9d97853543586596b86391f8de5d4f6c2a0e068"}, - {file = "aiohttp-3.12.12-cp39-cp39-win_amd64.whl", hash = "sha256:ddf40ba4a1d0b4d232dc47d2b98ae7e937dcbc40bb5f2746bce0af490a64526f"}, - {file = "aiohttp-3.12.12.tar.gz", hash = "sha256:05875595d2483d96cb61fa9f64e75262d7ac6251a7e3c811d8e26f7d721760bd"}, + {file = "aiohttp-3.12.14-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:906d5075b5ba0dd1c66fcaaf60eb09926a9fef3ca92d912d2a0bbdbecf8b1248"}, + {file = "aiohttp-3.12.14-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:c875bf6fc2fd1a572aba0e02ef4e7a63694778c5646cdbda346ee24e630d30fb"}, + {file = "aiohttp-3.12.14-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:fbb284d15c6a45fab030740049d03c0ecd60edad9cd23b211d7e11d3be8d56fd"}, + {file = "aiohttp-3.12.14-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:38e360381e02e1a05d36b223ecab7bc4a6e7b5ab15760022dc92589ee1d4238c"}, + {file = "aiohttp-3.12.14-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:aaf90137b5e5d84a53632ad95ebee5c9e3e7468f0aab92ba3f608adcb914fa95"}, + {file = "aiohttp-3.12.14-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:e532a25e4a0a2685fa295a31acf65e027fbe2bea7a4b02cdfbbba8a064577663"}, + {file = "aiohttp-3.12.14-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:eab9762c4d1b08ae04a6c77474e6136da722e34fdc0e6d6eab5ee93ac29f35d1"}, + {file = "aiohttp-3.12.14-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:abe53c3812b2899889a7fca763cdfaeee725f5be68ea89905e4275476ffd7e61"}, + {file = "aiohttp-3.12.14-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5760909b7080aa2ec1d320baee90d03b21745573780a072b66ce633eb77a8656"}, + {file = "aiohttp-3.12.14-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:02fcd3f69051467bbaa7f84d7ec3267478c7df18d68b2e28279116e29d18d4f3"}, + {file = "aiohttp-3.12.14-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:4dcd1172cd6794884c33e504d3da3c35648b8be9bfa946942d353b939d5f1288"}, + {file = "aiohttp-3.12.14-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:224d0da41355b942b43ad08101b1b41ce633a654128ee07e36d75133443adcda"}, + {file = "aiohttp-3.12.14-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:e387668724f4d734e865c1776d841ed75b300ee61059aca0b05bce67061dcacc"}, + {file = "aiohttp-3.12.14-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:dec9cde5b5a24171e0b0a4ca064b1414950904053fb77c707efd876a2da525d8"}, + {file = "aiohttp-3.12.14-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:bbad68a2af4877cc103cd94af9160e45676fc6f0c14abb88e6e092b945c2c8e3"}, + {file = "aiohttp-3.12.14-cp310-cp310-win32.whl", hash = "sha256:ee580cb7c00bd857b3039ebca03c4448e84700dc1322f860cf7a500a6f62630c"}, + {file = "aiohttp-3.12.14-cp310-cp310-win_amd64.whl", hash = "sha256:cf4f05b8cea571e2ccc3ca744e35ead24992d90a72ca2cf7ab7a2efbac6716db"}, + {file = "aiohttp-3.12.14-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:f4552ff7b18bcec18b60a90c6982049cdb9dac1dba48cf00b97934a06ce2e597"}, + {file = "aiohttp-3.12.14-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:8283f42181ff6ccbcf25acaae4e8ab2ff7e92b3ca4a4ced73b2c12d8cd971393"}, + {file = "aiohttp-3.12.14-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:040afa180ea514495aaff7ad34ec3d27826eaa5d19812730fe9e529b04bb2179"}, + {file = "aiohttp-3.12.14-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:b413c12f14c1149f0ffd890f4141a7471ba4b41234fe4fd4a0ff82b1dc299dbb"}, + {file = "aiohttp-3.12.14-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:1d6f607ce2e1a93315414e3d448b831238f1874b9968e1195b06efaa5c87e245"}, + {file = "aiohttp-3.12.14-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:565e70d03e924333004ed101599902bba09ebb14843c8ea39d657f037115201b"}, + {file = "aiohttp-3.12.14-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:4699979560728b168d5ab63c668a093c9570af2c7a78ea24ca5212c6cdc2b641"}, + {file = "aiohttp-3.12.14-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ad5fdf6af93ec6c99bf800eba3af9a43d8bfd66dce920ac905c817ef4a712afe"}, + {file = "aiohttp-3.12.14-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4ac76627c0b7ee0e80e871bde0d376a057916cb008a8f3ffc889570a838f5cc7"}, + {file = "aiohttp-3.12.14-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:798204af1180885651b77bf03adc903743a86a39c7392c472891649610844635"}, + {file = "aiohttp-3.12.14-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:4f1205f97de92c37dd71cf2d5bcfb65fdaed3c255d246172cce729a8d849b4da"}, + {file = "aiohttp-3.12.14-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:76ae6f1dd041f85065d9df77c6bc9c9703da9b5c018479d20262acc3df97d419"}, + {file = "aiohttp-3.12.14-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:a194ace7bc43ce765338ca2dfb5661489317db216ea7ea700b0332878b392cab"}, + {file = "aiohttp-3.12.14-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:16260e8e03744a6fe3fcb05259eeab8e08342c4c33decf96a9dad9f1187275d0"}, + {file = "aiohttp-3.12.14-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:8c779e5ebbf0e2e15334ea404fcce54009dc069210164a244d2eac8352a44b28"}, + {file = "aiohttp-3.12.14-cp311-cp311-win32.whl", hash = "sha256:a289f50bf1bd5be227376c067927f78079a7bdeccf8daa6a9e65c38bae14324b"}, + {file = "aiohttp-3.12.14-cp311-cp311-win_amd64.whl", hash = "sha256:0b8a69acaf06b17e9c54151a6c956339cf46db4ff72b3ac28516d0f7068f4ced"}, + {file = "aiohttp-3.12.14-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:a0ecbb32fc3e69bc25efcda7d28d38e987d007096cbbeed04f14a6662d0eee22"}, + {file = "aiohttp-3.12.14-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:0400f0ca9bb3e0b02f6466421f253797f6384e9845820c8b05e976398ac1d81a"}, + {file = "aiohttp-3.12.14-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:a56809fed4c8a830b5cae18454b7464e1529dbf66f71c4772e3cfa9cbec0a1ff"}, + {file = "aiohttp-3.12.14-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:27f2e373276e4755691a963e5d11756d093e346119f0627c2d6518208483fb6d"}, + {file = "aiohttp-3.12.14-cp312-cp312-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:ca39e433630e9a16281125ef57ece6817afd1d54c9f1bf32e901f38f16035869"}, + {file = "aiohttp-3.12.14-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9c748b3f8b14c77720132b2510a7d9907a03c20ba80f469e58d5dfd90c079a1c"}, + {file = "aiohttp-3.12.14-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f0a568abe1b15ce69d4cc37e23020720423f0728e3cb1f9bcd3f53420ec3bfe7"}, + {file = "aiohttp-3.12.14-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9888e60c2c54eaf56704b17feb558c7ed6b7439bca1e07d4818ab878f2083660"}, + {file = "aiohttp-3.12.14-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:3006a1dc579b9156de01e7916d38c63dc1ea0679b14627a37edf6151bc530088"}, + {file = "aiohttp-3.12.14-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:aa8ec5c15ab80e5501a26719eb48a55f3c567da45c6ea5bb78c52c036b2655c7"}, + {file = "aiohttp-3.12.14-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:39b94e50959aa07844c7fe2206b9f75d63cc3ad1c648aaa755aa257f6f2498a9"}, + {file = "aiohttp-3.12.14-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:04c11907492f416dad9885d503fbfc5dcb6768d90cad8639a771922d584609d3"}, + {file = "aiohttp-3.12.14-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:88167bd9ab69bb46cee91bd9761db6dfd45b6e76a0438c7e884c3f8160ff21eb"}, + {file = "aiohttp-3.12.14-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:791504763f25e8f9f251e4688195e8b455f8820274320204f7eafc467e609425"}, + {file = "aiohttp-3.12.14-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:2785b112346e435dd3a1a67f67713a3fe692d288542f1347ad255683f066d8e0"}, + {file = "aiohttp-3.12.14-cp312-cp312-win32.whl", hash = "sha256:15f5f4792c9c999a31d8decf444e79fcfd98497bf98e94284bf390a7bb8c1729"}, + {file = "aiohttp-3.12.14-cp312-cp312-win_amd64.whl", hash = "sha256:3b66e1a182879f579b105a80d5c4bd448b91a57e8933564bf41665064796a338"}, + {file = "aiohttp-3.12.14-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:3143a7893d94dc82bc409f7308bc10d60285a3cd831a68faf1aa0836c5c3c767"}, + {file = "aiohttp-3.12.14-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:3d62ac3d506cef54b355bd34c2a7c230eb693880001dfcda0bf88b38f5d7af7e"}, + {file = "aiohttp-3.12.14-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:48e43e075c6a438937c4de48ec30fa8ad8e6dfef122a038847456bfe7b947b63"}, + {file = "aiohttp-3.12.14-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:077b4488411a9724cecc436cbc8c133e0d61e694995b8de51aaf351c7578949d"}, + {file = "aiohttp-3.12.14-cp313-cp313-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:d8c35632575653f297dcbc9546305b2c1133391089ab925a6a3706dfa775ccab"}, + {file = "aiohttp-3.12.14-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:6b8ce87963f0035c6834b28f061df90cf525ff7c9b6283a8ac23acee6502afd4"}, + {file = "aiohttp-3.12.14-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f0a2cf66e32a2563bb0766eb24eae7e9a269ac0dc48db0aae90b575dc9583026"}, + {file = "aiohttp-3.12.14-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cdea089caf6d5cde975084a884c72d901e36ef9c2fd972c9f51efbbc64e96fbd"}, + {file = "aiohttp-3.12.14-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:8a7865f27db67d49e81d463da64a59365ebd6b826e0e4847aa111056dcb9dc88"}, + {file = "aiohttp-3.12.14-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:0ab5b38a6a39781d77713ad930cb5e7feea6f253de656a5f9f281a8f5931b086"}, + {file = "aiohttp-3.12.14-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:9b3b15acee5c17e8848d90a4ebc27853f37077ba6aec4d8cb4dbbea56d156933"}, + {file = "aiohttp-3.12.14-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:e4c972b0bdaac167c1e53e16a16101b17c6d0ed7eac178e653a07b9f7fad7151"}, + {file = "aiohttp-3.12.14-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:7442488b0039257a3bdbc55f7209587911f143fca11df9869578db6c26feeeb8"}, + {file = "aiohttp-3.12.14-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:f68d3067eecb64c5e9bab4a26aa11bd676f4c70eea9ef6536b0a4e490639add3"}, + {file = "aiohttp-3.12.14-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:f88d3704c8b3d598a08ad17d06006cb1ca52a1182291f04979e305c8be6c9758"}, + {file = "aiohttp-3.12.14-cp313-cp313-win32.whl", hash = "sha256:a3c99ab19c7bf375c4ae3debd91ca5d394b98b6089a03231d4c580ef3c2ae4c5"}, + {file = "aiohttp-3.12.14-cp313-cp313-win_amd64.whl", hash = "sha256:3f8aad695e12edc9d571f878c62bedc91adf30c760c8632f09663e5f564f4baa"}, + {file = "aiohttp-3.12.14-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:b8cc6b05e94d837bcd71c6531e2344e1ff0fb87abe4ad78a9261d67ef5d83eae"}, + {file = "aiohttp-3.12.14-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:d1dcb015ac6a3b8facd3677597edd5ff39d11d937456702f0bb2b762e390a21b"}, + {file = "aiohttp-3.12.14-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3779ed96105cd70ee5e85ca4f457adbce3d9ff33ec3d0ebcdf6c5727f26b21b3"}, + {file = "aiohttp-3.12.14-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:717a0680729b4ebd7569c1dcd718c46b09b360745fd8eb12317abc74b14d14d0"}, + {file = "aiohttp-3.12.14-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:b5dd3a2ef7c7e968dbbac8f5574ebeac4d2b813b247e8cec28174a2ba3627170"}, + {file = "aiohttp-3.12.14-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4710f77598c0092239bc12c1fcc278a444e16c7032d91babf5abbf7166463f7b"}, + {file = "aiohttp-3.12.14-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f3e9f75ae842a6c22a195d4a127263dbf87cbab729829e0bd7857fb1672400b2"}, + {file = "aiohttp-3.12.14-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5f9c8d55d6802086edd188e3a7d85a77787e50d56ce3eb4757a3205fa4657922"}, + {file = "aiohttp-3.12.14-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:79b29053ff3ad307880d94562cca80693c62062a098a5776ea8ef5ef4b28d140"}, + {file = "aiohttp-3.12.14-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:23e1332fff36bebd3183db0c7a547a1da9d3b4091509f6d818e098855f2f27d3"}, + {file = "aiohttp-3.12.14-cp39-cp39-musllinux_1_2_armv7l.whl", hash = "sha256:a564188ce831fd110ea76bcc97085dd6c625b427db3f1dbb14ca4baa1447dcbc"}, + {file = "aiohttp-3.12.14-cp39-cp39-musllinux_1_2_i686.whl", hash = "sha256:a7a1b4302f70bb3ec40ca86de82def532c97a80db49cac6a6700af0de41af5ee"}, + {file = "aiohttp-3.12.14-cp39-cp39-musllinux_1_2_ppc64le.whl", hash = "sha256:1b07ccef62950a2519f9bfc1e5b294de5dd84329f444ca0b329605ea787a3de5"}, + {file = "aiohttp-3.12.14-cp39-cp39-musllinux_1_2_s390x.whl", hash = "sha256:938bd3ca6259e7e48b38d84f753d548bd863e0c222ed6ee6ace3fd6752768a84"}, + {file = "aiohttp-3.12.14-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:8bc784302b6b9f163b54c4e93d7a6f09563bd01ff2b841b29ed3ac126e5040bf"}, + {file = "aiohttp-3.12.14-cp39-cp39-win32.whl", hash = "sha256:a3416f95961dd7d5393ecff99e3f41dc990fb72eda86c11f2a60308ac6dcd7a0"}, + {file = "aiohttp-3.12.14-cp39-cp39-win_amd64.whl", hash = "sha256:196858b8820d7f60578f8b47e5669b3195c21d8ab261e39b1d705346458f445f"}, + {file = "aiohttp-3.12.14.tar.gz", hash = "sha256:6e06e120e34d93100de448fd941522e11dafa78ef1a893c179901b7d66aa29f2"}, ] [package.dependencies] aiohappyeyeballs = ">=2.5.0" -aiosignal = ">=1.1.2" +aiosignal = ">=1.4.0" attrs = ">=17.3.0" frozenlist = ">=1.1.1" multidict = ">=4.5,<7.0" @@ -137,18 +137,19 @@ aiohttp = ">=3.9" [[package]] name = "aiosignal" -version = "1.3.2" +version = "1.4.0" description = "aiosignal: a list of registered asynchronous callbacks" optional = false python-versions = ">=3.9" groups = ["main"] files = [ - {file = "aiosignal-1.3.2-py2.py3-none-any.whl", hash = "sha256:45cde58e409a301715980c2b01d0c28bdde3770d8290b5eb2173759d9acb31a5"}, - {file = "aiosignal-1.3.2.tar.gz", hash = "sha256:a8c255c66fafb1e499c9351d0bf32ff2d8a0321595ebac3b93713656d2436f54"}, + {file = "aiosignal-1.4.0-py3-none-any.whl", hash = "sha256:053243f8b92b990551949e63930a839ff0cf0b0ebbe0597b0f3fb19e1a0fe82e"}, + {file = "aiosignal-1.4.0.tar.gz", hash = "sha256:f47eecd9468083c2029cc99945502cb7708b082c232f9aca65da147157b251c7"}, ] [package.dependencies] frozenlist = ">=1.1.0" +typing-extensions = {version = ">=4.2", markers = "python_version < \"3.13\""} [[package]] name = "alabaster" @@ -466,14 +467,14 @@ files = [ [[package]] name = "certifi" -version = "2025.4.26" +version = "2025.7.9" description = "Python package for providing Mozilla's CA Bundle." optional = false -python-versions = ">=3.6" +python-versions = ">=3.7" groups = ["main", "docs", "test"] files = [ - {file = "certifi-2025.4.26-py3-none-any.whl", hash = "sha256:30350364dfe371162649852c63336a15c70c6510c2ad5015b21c2345311805f3"}, - {file = "certifi-2025.4.26.tar.gz", hash = "sha256:0a816057ea3cdefcef70270d2c515e4506bbc954f417fa5ade2021213bb8f0c6"}, + {file = "certifi-2025.7.9-py3-none-any.whl", hash = "sha256:d842783a14f8fdd646895ac26f719a061408834473cfc10203f6a575beb15d39"}, + {file = "certifi-2025.7.9.tar.gz", hash = "sha256:c1d2ec05395148ee10cf672ffc28cd37ea0ab0d99f9cc74c43e588cbd111b079"}, ] [[package]] @@ -556,6 +557,18 @@ files = [ [package.dependencies] pycparser = "*" +[[package]] +name = "chardet" +version = "5.2.0" +description = "Universal encoding detector for Python 3" +optional = false +python-versions = ">=3.7" +groups = ["dev"] +files = [ + {file = "chardet-5.2.0-py3-none-any.whl", hash = "sha256:e1cf59446890a00105fe7b7912492ea04b6e6f06d4b742b2c788469e34c82970"}, + {file = "chardet-5.2.0.tar.gz", hash = "sha256:1b3b6ff479a8c414bc3fa2c0852995695c4a026dcd6d0633b2dd092ca39c1cf7"}, +] + [[package]] name = "charset-normalizer" version = "3.4.2" @@ -688,14 +701,14 @@ markers = {main = "platform_system == \"Windows\" or sys_platform == \"win32\"", [[package]] name = "colorful" -version = "0.5.6" +version = "0.5.7" description = "Terminal string styling done right, in Python." optional = false python-versions = "*" groups = ["main"] files = [ - {file = "colorful-0.5.6-py2.py3-none-any.whl", hash = "sha256:eab8c1c809f5025ad2b5238a50bd691e26850da8cac8f90d660ede6ea1af9f1e"}, - {file = "colorful-0.5.6.tar.gz", hash = "sha256:b56d5c01db1dac4898308ea889edcb113fbee3e6ec5df4bacffd61d5241b5b8d"}, + {file = "colorful-0.5.7-py2.py3-none-any.whl", hash = "sha256:495dd3a23151a9568cee8a90fc1174c902ad7ef06655f50b6bddf9e80008da69"}, + {file = "colorful-0.5.7.tar.gz", hash = "sha256:c5452179b56601c178b03d468a5326cc1fe37d9be81d24d0d6bdab36c4b93ad8"}, ] [package.dependencies] @@ -904,6 +917,27 @@ files = [ {file = "defusedxml-0.7.1.tar.gz", hash = "sha256:1bb3032db185915b62d7c6209c5a8792be6a32ab2fedacc84e01b52c51aa3e69"}, ] +[[package]] +name = "diff-cover" +version = "9.6.0" +description = "Run coverage and linting reports on diffs" +optional = false +python-versions = ">=3.9" +groups = ["dev"] +files = [ + {file = "diff_cover-9.6.0-py3-none-any.whl", hash = "sha256:29fbeb52d77a0b8c811e5580d5dbf41801a838da2ed54319a599da8f7233c547"}, + {file = "diff_cover-9.6.0.tar.gz", hash = "sha256:75e5bc056dcaa68c6c87c9fb4e07c9e60daef15b6e8d034d56d2da9e2c84a872"}, +] + +[package.dependencies] +chardet = ">=3.0.0" +Jinja2 = ">=2.7.1" +pluggy = ">=0.13.1,<2" +Pygments = ">=2.19.1,<3.0.0" + +[package.extras] +toml = ["tomli (>=1.2.1)"] + [[package]] name = "distlib" version = "0.3.9" @@ -1155,14 +1189,14 @@ tqdm = ["tqdm"] [[package]] name = "google-api-core" -version = "2.25.0" +version = "2.25.1" description = "Google API client core library" optional = false python-versions = ">=3.7" groups = ["main"] files = [ - {file = "google_api_core-2.25.0-py3-none-any.whl", hash = "sha256:1db79d1281dcf9f3d10023283299ba38f3dc9f639ec41085968fd23e5bcf512e"}, - {file = "google_api_core-2.25.0.tar.gz", hash = "sha256:9b548e688702f82a34ed8409fb8a6961166f0b7795032f0be8f48308dff4333a"}, + {file = "google_api_core-2.25.1-py3-none-any.whl", hash = "sha256:8a2a56c1fef82987a524371f99f3bd0143702fecc670c72e600c1cda6bf8dbb7"}, + {file = "google_api_core-2.25.1.tar.gz", hash = "sha256:d2aaa0b13c78c61cb3f4282c464c046e45fbd75755683c9c525e6e8f7ed0a5e8"}, ] [package.dependencies] @@ -1228,67 +1262,67 @@ grpc = ["grpcio (>=1.44.0,<2.0.0)"] [[package]] name = "grpcio" -version = "1.73.0" +version = "1.73.1" description = "HTTP/2-based RPC framework" optional = false python-versions = ">=3.9" groups = ["main"] files = [ - {file = "grpcio-1.73.0-cp310-cp310-linux_armv7l.whl", hash = "sha256:d050197eeed50f858ef6c51ab09514856f957dba7b1f7812698260fc9cc417f6"}, - {file = "grpcio-1.73.0-cp310-cp310-macosx_11_0_universal2.whl", hash = "sha256:ebb8d5f4b0200916fb292a964a4d41210de92aba9007e33d8551d85800ea16cb"}, - {file = "grpcio-1.73.0-cp310-cp310-manylinux_2_17_aarch64.whl", hash = "sha256:c0811331b469e3f15dda5f90ab71bcd9681189a83944fd6dc908e2c9249041ef"}, - {file = "grpcio-1.73.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:12787c791c3993d0ea1cc8bf90393647e9a586066b3b322949365d2772ba965b"}, - {file = "grpcio-1.73.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2c17771e884fddf152f2a0df12478e8d02853e5b602a10a9a9f1f52fa02b1d32"}, - {file = "grpcio-1.73.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:275e23d4c428c26b51857bbd95fcb8e528783597207ec592571e4372b300a29f"}, - {file = "grpcio-1.73.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:9ffc972b530bf73ef0f948f799482a1bf12d9b6f33406a8e6387c0ca2098a833"}, - {file = "grpcio-1.73.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:ebd8d269df64aff092b2cec5e015d8ae09c7e90888b5c35c24fdca719a2c9f35"}, - {file = "grpcio-1.73.0-cp310-cp310-win32.whl", hash = "sha256:072d8154b8f74300ed362c01d54af8b93200c1a9077aeaea79828d48598514f1"}, - {file = "grpcio-1.73.0-cp310-cp310-win_amd64.whl", hash = "sha256:ce953d9d2100e1078a76a9dc2b7338d5415924dc59c69a15bf6e734db8a0f1ca"}, - {file = "grpcio-1.73.0-cp311-cp311-linux_armv7l.whl", hash = "sha256:51036f641f171eebe5fa7aaca5abbd6150f0c338dab3a58f9111354240fe36ec"}, - {file = "grpcio-1.73.0-cp311-cp311-macosx_11_0_universal2.whl", hash = "sha256:d12bbb88381ea00bdd92c55aff3da3391fd85bc902c41275c8447b86f036ce0f"}, - {file = "grpcio-1.73.0-cp311-cp311-manylinux_2_17_aarch64.whl", hash = "sha256:483c507c2328ed0e01bc1adb13d1eada05cc737ec301d8e5a8f4a90f387f1790"}, - {file = "grpcio-1.73.0-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c201a34aa960c962d0ce23fe5f423f97e9d4b518ad605eae6d0a82171809caaa"}, - {file = "grpcio-1.73.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:859f70c8e435e8e1fa060e04297c6818ffc81ca9ebd4940e180490958229a45a"}, - {file = "grpcio-1.73.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:e2459a27c6886e7e687e4e407778425f3c6a971fa17a16420227bda39574d64b"}, - {file = "grpcio-1.73.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:e0084d4559ee3dbdcce9395e1bc90fdd0262529b32c417a39ecbc18da8074ac7"}, - {file = "grpcio-1.73.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:ef5fff73d5f724755693a464d444ee0a448c6cdfd3c1616a9223f736c622617d"}, - {file = "grpcio-1.73.0-cp311-cp311-win32.whl", hash = "sha256:965a16b71a8eeef91fc4df1dc40dc39c344887249174053814f8a8e18449c4c3"}, - {file = "grpcio-1.73.0-cp311-cp311-win_amd64.whl", hash = "sha256:b71a7b4483d1f753bbc11089ff0f6fa63b49c97a9cc20552cded3fcad466d23b"}, - {file = "grpcio-1.73.0-cp312-cp312-linux_armv7l.whl", hash = "sha256:fb9d7c27089d9ba3746f18d2109eb530ef2a37452d2ff50f5a6696cd39167d3b"}, - {file = "grpcio-1.73.0-cp312-cp312-macosx_11_0_universal2.whl", hash = "sha256:128ba2ebdac41e41554d492b82c34586a90ebd0766f8ebd72160c0e3a57b9155"}, - {file = "grpcio-1.73.0-cp312-cp312-manylinux_2_17_aarch64.whl", hash = "sha256:068ecc415f79408d57a7f146f54cdf9f0acb4b301a52a9e563973dc981e82f3d"}, - {file = "grpcio-1.73.0-cp312-cp312-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6ddc1cfb2240f84d35d559ade18f69dcd4257dbaa5ba0de1a565d903aaab2968"}, - {file = "grpcio-1.73.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e53007f70d9783f53b41b4cf38ed39a8e348011437e4c287eee7dd1d39d54b2f"}, - {file = "grpcio-1.73.0-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:4dd8d8d092efede7d6f48d695ba2592046acd04ccf421436dd7ed52677a9ad29"}, - {file = "grpcio-1.73.0-cp312-cp312-musllinux_1_1_i686.whl", hash = "sha256:70176093d0a95b44d24baa9c034bb67bfe2b6b5f7ebc2836f4093c97010e17fd"}, - {file = "grpcio-1.73.0-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:085ebe876373ca095e24ced95c8f440495ed0b574c491f7f4f714ff794bbcd10"}, - {file = "grpcio-1.73.0-cp312-cp312-win32.whl", hash = "sha256:cfc556c1d6aef02c727ec7d0016827a73bfe67193e47c546f7cadd3ee6bf1a60"}, - {file = "grpcio-1.73.0-cp312-cp312-win_amd64.whl", hash = "sha256:bbf45d59d090bf69f1e4e1594832aaf40aa84b31659af3c5e2c3f6a35202791a"}, - {file = "grpcio-1.73.0-cp313-cp313-linux_armv7l.whl", hash = "sha256:da1d677018ef423202aca6d73a8d3b2cb245699eb7f50eb5f74cae15a8e1f724"}, - {file = "grpcio-1.73.0-cp313-cp313-macosx_11_0_universal2.whl", hash = "sha256:36bf93f6a657f37c131d9dd2c391b867abf1426a86727c3575393e9e11dadb0d"}, - {file = "grpcio-1.73.0-cp313-cp313-manylinux_2_17_aarch64.whl", hash = "sha256:d84000367508ade791d90c2bafbd905574b5ced8056397027a77a215d601ba15"}, - {file = "grpcio-1.73.0-cp313-cp313-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:c98ba1d928a178ce33f3425ff823318040a2b7ef875d30a0073565e5ceb058d9"}, - {file = "grpcio-1.73.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a73c72922dfd30b396a5f25bb3a4590195ee45ecde7ee068acb0892d2900cf07"}, - {file = "grpcio-1.73.0-cp313-cp313-musllinux_1_1_aarch64.whl", hash = "sha256:10e8edc035724aba0346a432060fd192b42bd03675d083c01553cab071a28da5"}, - {file = "grpcio-1.73.0-cp313-cp313-musllinux_1_1_i686.whl", hash = "sha256:f5cdc332b503c33b1643b12ea933582c7b081957c8bc2ea4cc4bc58054a09288"}, - {file = "grpcio-1.73.0-cp313-cp313-musllinux_1_1_x86_64.whl", hash = "sha256:07ad7c57233c2109e4ac999cb9c2710c3b8e3f491a73b058b0ce431f31ed8145"}, - {file = "grpcio-1.73.0-cp313-cp313-win32.whl", hash = "sha256:0eb5df4f41ea10bda99a802b2a292d85be28958ede2a50f2beb8c7fc9a738419"}, - {file = "grpcio-1.73.0-cp313-cp313-win_amd64.whl", hash = "sha256:38cf518cc54cd0c47c9539cefa8888549fcc067db0b0c66a46535ca8032020c4"}, - {file = "grpcio-1.73.0-cp39-cp39-linux_armv7l.whl", hash = "sha256:1284850607901cfe1475852d808e5a102133461ec9380bc3fc9ebc0686ee8e32"}, - {file = "grpcio-1.73.0-cp39-cp39-macosx_11_0_universal2.whl", hash = "sha256:0e092a4b28eefb63eec00d09ef33291cd4c3a0875cde29aec4d11d74434d222c"}, - {file = "grpcio-1.73.0-cp39-cp39-manylinux_2_17_aarch64.whl", hash = "sha256:33577fe7febffe8ebad458744cfee8914e0c10b09f0ff073a6b149a84df8ab8f"}, - {file = "grpcio-1.73.0-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:60813d8a16420d01fa0da1fc7ebfaaa49a7e5051b0337cd48f4f950eb249a08e"}, - {file = "grpcio-1.73.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2a9c957dc65e5d474378d7bcc557e9184576605d4b4539e8ead6e351d7ccce20"}, - {file = "grpcio-1.73.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:3902b71407d021163ea93c70c8531551f71ae742db15b66826cf8825707d2908"}, - {file = "grpcio-1.73.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:1dd7fa7276dcf061e2d5f9316604499eea06b1b23e34a9380572d74fe59915a8"}, - {file = "grpcio-1.73.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:2d1510c4ea473110cb46a010555f2c1a279d1c256edb276e17fa571ba1e8927c"}, - {file = "grpcio-1.73.0-cp39-cp39-win32.whl", hash = "sha256:d0a1517b2005ba1235a1190b98509264bf72e231215dfeef8db9a5a92868789e"}, - {file = "grpcio-1.73.0-cp39-cp39-win_amd64.whl", hash = "sha256:6228f7eb6d9f785f38b589d49957fca5df3d5b5349e77d2d89b14e390165344c"}, - {file = "grpcio-1.73.0.tar.gz", hash = "sha256:3af4c30918a7f0d39de500d11255f8d9da4f30e94a2033e70fe2a720e184bd8e"}, + {file = "grpcio-1.73.1-cp310-cp310-linux_armv7l.whl", hash = "sha256:2d70f4ddd0a823436c2624640570ed6097e40935c9194482475fe8e3d9754d55"}, + {file = "grpcio-1.73.1-cp310-cp310-macosx_11_0_universal2.whl", hash = "sha256:3841a8a5a66830261ab6a3c2a3dc539ed84e4ab019165f77b3eeb9f0ba621f26"}, + {file = "grpcio-1.73.1-cp310-cp310-manylinux_2_17_aarch64.whl", hash = "sha256:628c30f8e77e0258ab788750ec92059fc3d6628590fb4b7cea8c102503623ed7"}, + {file = "grpcio-1.73.1-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:67a0468256c9db6d5ecb1fde4bf409d016f42cef649323f0a08a72f352d1358b"}, + {file = "grpcio-1.73.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:68b84d65bbdebd5926eb5c53b0b9ec3b3f83408a30e4c20c373c5337b4219ec5"}, + {file = "grpcio-1.73.1-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:c54796ca22b8349cc594d18b01099e39f2b7ffb586ad83217655781a350ce4da"}, + {file = "grpcio-1.73.1-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:75fc8e543962ece2f7ecd32ada2d44c0c8570ae73ec92869f9af8b944863116d"}, + {file = "grpcio-1.73.1-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:6a6037891cd2b1dd1406b388660522e1565ed340b1fea2955b0234bdd941a862"}, + {file = "grpcio-1.73.1-cp310-cp310-win32.whl", hash = "sha256:cce7265b9617168c2d08ae570fcc2af4eaf72e84f8c710ca657cc546115263af"}, + {file = "grpcio-1.73.1-cp310-cp310-win_amd64.whl", hash = "sha256:6a2b372e65fad38842050943f42ce8fee00c6f2e8ea4f7754ba7478d26a356ee"}, + {file = "grpcio-1.73.1-cp311-cp311-linux_armv7l.whl", hash = "sha256:ba2cea9f7ae4bc21f42015f0ec98f69ae4179848ad744b210e7685112fa507a1"}, + {file = "grpcio-1.73.1-cp311-cp311-macosx_11_0_universal2.whl", hash = "sha256:d74c3f4f37b79e746271aa6cdb3a1d7e4432aea38735542b23adcabaaee0c097"}, + {file = "grpcio-1.73.1-cp311-cp311-manylinux_2_17_aarch64.whl", hash = "sha256:5b9b1805a7d61c9e90541cbe8dfe0a593dfc8c5c3a43fe623701b6a01b01d710"}, + {file = "grpcio-1.73.1-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:b3215f69a0670a8cfa2ab53236d9e8026bfb7ead5d4baabe7d7dc11d30fda967"}, + {file = "grpcio-1.73.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bc5eccfd9577a5dc7d5612b2ba90cca4ad14c6d949216c68585fdec9848befb1"}, + {file = "grpcio-1.73.1-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:dc7d7fd520614fce2e6455ba89791458020a39716951c7c07694f9dbae28e9c0"}, + {file = "grpcio-1.73.1-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:105492124828911f85127e4825d1c1234b032cb9d238567876b5515d01151379"}, + {file = "grpcio-1.73.1-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:610e19b04f452ba6f402ac9aa94eb3d21fbc94553368008af634812c4a85a99e"}, + {file = "grpcio-1.73.1-cp311-cp311-win32.whl", hash = "sha256:d60588ab6ba0ac753761ee0e5b30a29398306401bfbceffe7d68ebb21193f9d4"}, + {file = "grpcio-1.73.1-cp311-cp311-win_amd64.whl", hash = "sha256:6957025a4608bb0a5ff42abd75bfbb2ed99eda29d5992ef31d691ab54b753643"}, + {file = "grpcio-1.73.1-cp312-cp312-linux_armv7l.whl", hash = "sha256:921b25618b084e75d424a9f8e6403bfeb7abef074bb6c3174701e0f2542debcf"}, + {file = "grpcio-1.73.1-cp312-cp312-macosx_11_0_universal2.whl", hash = "sha256:277b426a0ed341e8447fbf6c1d6b68c952adddf585ea4685aa563de0f03df887"}, + {file = "grpcio-1.73.1-cp312-cp312-manylinux_2_17_aarch64.whl", hash = "sha256:96c112333309493c10e118d92f04594f9055774757f5d101b39f8150f8c25582"}, + {file = "grpcio-1.73.1-cp312-cp312-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f48e862aed925ae987eb7084409a80985de75243389dc9d9c271dd711e589918"}, + {file = "grpcio-1.73.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:83a6c2cce218e28f5040429835fa34a29319071079e3169f9543c3fbeff166d2"}, + {file = "grpcio-1.73.1-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:65b0458a10b100d815a8426b1442bd17001fdb77ea13665b2f7dc9e8587fdc6b"}, + {file = "grpcio-1.73.1-cp312-cp312-musllinux_1_1_i686.whl", hash = "sha256:0a9f3ea8dce9eae9d7cb36827200133a72b37a63896e0e61a9d5ec7d61a59ab1"}, + {file = "grpcio-1.73.1-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:de18769aea47f18e782bf6819a37c1c528914bfd5683b8782b9da356506190c8"}, + {file = "grpcio-1.73.1-cp312-cp312-win32.whl", hash = "sha256:24e06a5319e33041e322d32c62b1e728f18ab8c9dbc91729a3d9f9e3ed336642"}, + {file = "grpcio-1.73.1-cp312-cp312-win_amd64.whl", hash = "sha256:303c8135d8ab176f8038c14cc10d698ae1db9c480f2b2823f7a987aa2a4c5646"}, + {file = "grpcio-1.73.1-cp313-cp313-linux_armv7l.whl", hash = "sha256:b310824ab5092cf74750ebd8a8a8981c1810cb2b363210e70d06ef37ad80d4f9"}, + {file = "grpcio-1.73.1-cp313-cp313-macosx_11_0_universal2.whl", hash = "sha256:8f5a6df3fba31a3485096ac85b2e34b9666ffb0590df0cd044f58694e6a1f6b5"}, + {file = "grpcio-1.73.1-cp313-cp313-manylinux_2_17_aarch64.whl", hash = "sha256:052e28fe9c41357da42250a91926a3e2f74c046575c070b69659467ca5aa976b"}, + {file = "grpcio-1.73.1-cp313-cp313-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1c0bf15f629b1497436596b1cbddddfa3234273490229ca29561209778ebe182"}, + {file = "grpcio-1.73.1-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0ab860d5bfa788c5a021fba264802e2593688cd965d1374d31d2b1a34cacd854"}, + {file = "grpcio-1.73.1-cp313-cp313-musllinux_1_1_aarch64.whl", hash = "sha256:ad1d958c31cc91ab050bd8a91355480b8e0683e21176522bacea225ce51163f2"}, + {file = "grpcio-1.73.1-cp313-cp313-musllinux_1_1_i686.whl", hash = "sha256:f43ffb3bd415c57224c7427bfb9e6c46a0b6e998754bfa0d00f408e1873dcbb5"}, + {file = "grpcio-1.73.1-cp313-cp313-musllinux_1_1_x86_64.whl", hash = "sha256:686231cdd03a8a8055f798b2b54b19428cdf18fa1549bee92249b43607c42668"}, + {file = "grpcio-1.73.1-cp313-cp313-win32.whl", hash = "sha256:89018866a096e2ce21e05eabed1567479713ebe57b1db7cbb0f1e3b896793ba4"}, + {file = "grpcio-1.73.1-cp313-cp313-win_amd64.whl", hash = "sha256:4a68f8c9966b94dff693670a5cf2b54888a48a5011c5d9ce2295a1a1465ee84f"}, + {file = "grpcio-1.73.1-cp39-cp39-linux_armv7l.whl", hash = "sha256:b4adc97d2d7f5c660a5498bda978ebb866066ad10097265a5da0511323ae9f50"}, + {file = "grpcio-1.73.1-cp39-cp39-macosx_11_0_universal2.whl", hash = "sha256:c45a28a0cfb6ddcc7dc50a29de44ecac53d115c3388b2782404218db51cb2df3"}, + {file = "grpcio-1.73.1-cp39-cp39-manylinux_2_17_aarch64.whl", hash = "sha256:10af9f2ab98a39f5b6c1896c6fc2036744b5b41d12739d48bed4c3e15b6cf900"}, + {file = "grpcio-1.73.1-cp39-cp39-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:45cf17dcce5ebdb7b4fe9e86cb338fa99d7d1bb71defc78228e1ddf8d0de8cbb"}, + {file = "grpcio-1.73.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1c502c2e950fc7e8bf05c047e8a14522ef7babac59abbfde6dbf46b7a0d9c71e"}, + {file = "grpcio-1.73.1-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:6abfc0f9153dc4924536f40336f88bd4fe7bd7494f028675e2e04291b8c2c62a"}, + {file = "grpcio-1.73.1-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:ed451a0e39c8e51eb1612b78686839efd1a920666d1666c1adfdb4fd51680c0f"}, + {file = "grpcio-1.73.1-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:07f08705a5505c9b5b0cbcbabafb96462b5a15b7236bbf6bbcc6b0b91e1cbd7e"}, + {file = "grpcio-1.73.1-cp39-cp39-win32.whl", hash = "sha256:ad5c958cc3d98bb9d71714dc69f1c13aaf2f4b53e29d4cc3f1501ef2e4d129b2"}, + {file = "grpcio-1.73.1-cp39-cp39-win_amd64.whl", hash = "sha256:42f0660bce31b745eb9d23f094a332d31f210dcadd0fc8e5be7e4c62a87ce86b"}, + {file = "grpcio-1.73.1.tar.gz", hash = "sha256:7fce2cd1c0c1116cf3850564ebfc3264fba75d3c74a7414373f1238ea365ef87"}, ] [package.extras] -protobuf = ["grpcio-tools (>=1.73.0)"] +protobuf = ["grpcio-tools (>=1.73.1)"] [[package]] name = "h11" @@ -1448,14 +1482,14 @@ test = ["flaky", "ipyparallel", "pre-commit", "pytest (>=7.0)", "pytest-asyncio [[package]] name = "ipython" -version = "9.3.0" +version = "9.4.0" description = "IPython: Productive Interactive Computing" optional = false python-versions = ">=3.11" groups = ["main", "test"] files = [ - {file = "ipython-9.3.0-py3-none-any.whl", hash = "sha256:1a0b6dd9221a1f5dddf725b57ac0cb6fddc7b5f470576231ae9162b9b3455a04"}, - {file = "ipython-9.3.0.tar.gz", hash = "sha256:79eb896f9f23f50ad16c3bc205f686f6e030ad246cc309c6279a242b14afe9d8"}, + {file = "ipython-9.4.0-py3-none-any.whl", hash = "sha256:25850f025a446d9b359e8d296ba175a36aedd32e83ca9b5060430fe16801f066"}, + {file = "ipython-9.4.0.tar.gz", hash = "sha256:c033c6d4e7914c3d9768aabe76bbe87ba1dc66a92a05db6bfa1125d81f2ee270"}, ] [package.dependencies] @@ -1557,7 +1591,7 @@ version = "3.1.6" description = "A very fast and expressive template engine." optional = false python-versions = ">=3.7" -groups = ["docs", "test"] +groups = ["dev", "docs", "test"] files = [ {file = "jinja2-3.1.6-py3-none-any.whl", hash = "sha256:85ece4451f492d0c13c5dd7c13a64681a86afae63a5f347908daf103ce6d2f67"}, {file = "jinja2-3.1.6.tar.gz", hash = "sha256:0137fb05990d35f1275a587e9aee6d56da821fc83491a0fb838183be43f66d6d"}, @@ -1927,7 +1961,7 @@ version = "3.0.2" description = "Safely add untrusted strings to HTML/XML markup." optional = false python-versions = ">=3.9" -groups = ["docs", "test"] +groups = ["dev", "docs", "test"] files = [ {file = "MarkupSafe-3.0.2-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:7e94c425039cde14257288fd61dcfb01963e658efbc0ff54f5306b06054700f8"}, {file = "MarkupSafe-3.0.2-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:9e2d922824181480953426608b81967de705c3cef4d1af983af849d7bd619158"}, @@ -2033,190 +2067,191 @@ files = [ [[package]] name = "msgpack" -version = "1.1.0" +version = "1.1.1" description = "MessagePack serializer" optional = false python-versions = ">=3.8" groups = ["main"] files = [ - {file = "msgpack-1.1.0-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:7ad442d527a7e358a469faf43fda45aaf4ac3249c8310a82f0ccff9164e5dccd"}, - {file = "msgpack-1.1.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:74bed8f63f8f14d75eec75cf3d04ad581da6b914001b474a5d3cd3372c8cc27d"}, - {file = "msgpack-1.1.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:914571a2a5b4e7606997e169f64ce53a8b1e06f2cf2c3a7273aa106236d43dd5"}, - {file = "msgpack-1.1.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c921af52214dcbb75e6bdf6a661b23c3e6417f00c603dd2070bccb5c3ef499f5"}, - {file = "msgpack-1.1.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d8ce0b22b890be5d252de90d0e0d119f363012027cf256185fc3d474c44b1b9e"}, - {file = "msgpack-1.1.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:73322a6cc57fcee3c0c57c4463d828e9428275fb85a27aa2aa1a92fdc42afd7b"}, - {file = "msgpack-1.1.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:e1f3c3d21f7cf67bcf2da8e494d30a75e4cf60041d98b3f79875afb5b96f3a3f"}, - {file = "msgpack-1.1.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:64fc9068d701233effd61b19efb1485587560b66fe57b3e50d29c5d78e7fef68"}, - {file = "msgpack-1.1.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:42f754515e0f683f9c79210a5d1cad631ec3d06cea5172214d2176a42e67e19b"}, - {file = "msgpack-1.1.0-cp310-cp310-win32.whl", hash = "sha256:3df7e6b05571b3814361e8464f9304c42d2196808e0119f55d0d3e62cd5ea044"}, - {file = "msgpack-1.1.0-cp310-cp310-win_amd64.whl", hash = "sha256:685ec345eefc757a7c8af44a3032734a739f8c45d1b0ac45efc5d8977aa4720f"}, - {file = "msgpack-1.1.0-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:3d364a55082fb2a7416f6c63ae383fbd903adb5a6cf78c5b96cc6316dc1cedc7"}, - {file = "msgpack-1.1.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:79ec007767b9b56860e0372085f8504db5d06bd6a327a335449508bbee9648fa"}, - {file = "msgpack-1.1.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:6ad622bf7756d5a497d5b6836e7fc3752e2dd6f4c648e24b1803f6048596f701"}, - {file = "msgpack-1.1.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8e59bca908d9ca0de3dc8684f21ebf9a690fe47b6be93236eb40b99af28b6ea6"}, - {file = "msgpack-1.1.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:5e1da8f11a3dd397f0a32c76165cf0c4eb95b31013a94f6ecc0b280c05c91b59"}, - {file = "msgpack-1.1.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:452aff037287acb1d70a804ffd022b21fa2bb7c46bee884dbc864cc9024128a0"}, - {file = "msgpack-1.1.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:8da4bf6d54ceed70e8861f833f83ce0814a2b72102e890cbdfe4b34764cdd66e"}, - {file = "msgpack-1.1.0-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:41c991beebf175faf352fb940bf2af9ad1fb77fd25f38d9142053914947cdbf6"}, - {file = "msgpack-1.1.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:a52a1f3a5af7ba1c9ace055b659189f6c669cf3657095b50f9602af3a3ba0fe5"}, - {file = "msgpack-1.1.0-cp311-cp311-win32.whl", hash = "sha256:58638690ebd0a06427c5fe1a227bb6b8b9fdc2bd07701bec13c2335c82131a88"}, - {file = "msgpack-1.1.0-cp311-cp311-win_amd64.whl", hash = "sha256:fd2906780f25c8ed5d7b323379f6138524ba793428db5d0e9d226d3fa6aa1788"}, - {file = "msgpack-1.1.0-cp312-cp312-macosx_10_9_universal2.whl", hash = "sha256:d46cf9e3705ea9485687aa4001a76e44748b609d260af21c4ceea7f2212a501d"}, - {file = "msgpack-1.1.0-cp312-cp312-macosx_10_9_x86_64.whl", hash = "sha256:5dbad74103df937e1325cc4bfeaf57713be0b4f15e1c2da43ccdd836393e2ea2"}, - {file = "msgpack-1.1.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:58dfc47f8b102da61e8949708b3eafc3504509a5728f8b4ddef84bd9e16ad420"}, - {file = "msgpack-1.1.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4676e5be1b472909b2ee6356ff425ebedf5142427842aa06b4dfd5117d1ca8a2"}, - {file = "msgpack-1.1.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:17fb65dd0bec285907f68b15734a993ad3fc94332b5bb21b0435846228de1f39"}, - {file = "msgpack-1.1.0-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:a51abd48c6d8ac89e0cfd4fe177c61481aca2d5e7ba42044fd218cfd8ea9899f"}, - {file = "msgpack-1.1.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:2137773500afa5494a61b1208619e3871f75f27b03bcfca7b3a7023284140247"}, - {file = "msgpack-1.1.0-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:398b713459fea610861c8a7b62a6fec1882759f308ae0795b5413ff6a160cf3c"}, - {file = "msgpack-1.1.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:06f5fd2f6bb2a7914922d935d3b8bb4a7fff3a9a91cfce6d06c13bc42bec975b"}, - {file = "msgpack-1.1.0-cp312-cp312-win32.whl", hash = "sha256:ad33e8400e4ec17ba782f7b9cf868977d867ed784a1f5f2ab46e7ba53b6e1e1b"}, - {file = "msgpack-1.1.0-cp312-cp312-win_amd64.whl", hash = "sha256:115a7af8ee9e8cddc10f87636767857e7e3717b7a2e97379dc2054712693e90f"}, - {file = "msgpack-1.1.0-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:071603e2f0771c45ad9bc65719291c568d4edf120b44eb36324dcb02a13bfddf"}, - {file = "msgpack-1.1.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:0f92a83b84e7c0749e3f12821949d79485971f087604178026085f60ce109330"}, - {file = "msgpack-1.1.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:4a1964df7b81285d00a84da4e70cb1383f2e665e0f1f2a7027e683956d04b734"}, - {file = "msgpack-1.1.0-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:59caf6a4ed0d164055ccff8fe31eddc0ebc07cf7326a2aaa0dbf7a4001cd823e"}, - {file = "msgpack-1.1.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0907e1a7119b337971a689153665764adc34e89175f9a34793307d9def08e6ca"}, - {file = "msgpack-1.1.0-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:65553c9b6da8166e819a6aa90ad15288599b340f91d18f60b2061f402b9a4915"}, - {file = "msgpack-1.1.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:7a946a8992941fea80ed4beae6bff74ffd7ee129a90b4dd5cf9c476a30e9708d"}, - {file = "msgpack-1.1.0-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:4b51405e36e075193bc051315dbf29168d6141ae2500ba8cd80a522964e31434"}, - {file = "msgpack-1.1.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:b4c01941fd2ff87c2a934ee6055bda4ed353a7846b8d4f341c428109e9fcde8c"}, - {file = "msgpack-1.1.0-cp313-cp313-win32.whl", hash = "sha256:7c9a35ce2c2573bada929e0b7b3576de647b0defbd25f5139dcdaba0ae35a4cc"}, - {file = "msgpack-1.1.0-cp313-cp313-win_amd64.whl", hash = "sha256:bce7d9e614a04d0883af0b3d4d501171fbfca038f12c77fa838d9f198147a23f"}, - {file = "msgpack-1.1.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:c40ffa9a15d74e05ba1fe2681ea33b9caffd886675412612d93ab17b58ea2fec"}, - {file = "msgpack-1.1.0-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f1ba6136e650898082d9d5a5217d5906d1e138024f836ff48691784bbe1adf96"}, - {file = "msgpack-1.1.0-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e0856a2b7e8dcb874be44fea031d22e5b3a19121be92a1e098f46068a11b0870"}, - {file = "msgpack-1.1.0-cp38-cp38-musllinux_1_2_aarch64.whl", hash = "sha256:471e27a5787a2e3f974ba023f9e265a8c7cfd373632247deb225617e3100a3c7"}, - {file = "msgpack-1.1.0-cp38-cp38-musllinux_1_2_i686.whl", hash = "sha256:646afc8102935a388ffc3914b336d22d1c2d6209c773f3eb5dd4d6d3b6f8c1cb"}, - {file = "msgpack-1.1.0-cp38-cp38-musllinux_1_2_x86_64.whl", hash = "sha256:13599f8829cfbe0158f6456374e9eea9f44eee08076291771d8ae93eda56607f"}, - {file = "msgpack-1.1.0-cp38-cp38-win32.whl", hash = "sha256:8a84efb768fb968381e525eeeb3d92857e4985aacc39f3c47ffd00eb4509315b"}, - {file = "msgpack-1.1.0-cp38-cp38-win_amd64.whl", hash = "sha256:879a7b7b0ad82481c52d3c7eb99bf6f0645dbdec5134a4bddbd16f3506947feb"}, - {file = "msgpack-1.1.0-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:53258eeb7a80fc46f62fd59c876957a2d0e15e6449a9e71842b6d24419d88ca1"}, - {file = "msgpack-1.1.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:7e7b853bbc44fb03fbdba34feb4bd414322180135e2cb5164f20ce1c9795ee48"}, - {file = "msgpack-1.1.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:f3e9b4936df53b970513eac1758f3882c88658a220b58dcc1e39606dccaaf01c"}, - {file = "msgpack-1.1.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:46c34e99110762a76e3911fc923222472c9d681f1094096ac4102c18319e6468"}, - {file = "msgpack-1.1.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8a706d1e74dd3dea05cb54580d9bd8b2880e9264856ce5068027eed09680aa74"}, - {file = "msgpack-1.1.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:534480ee5690ab3cbed89d4c8971a5c631b69a8c0883ecfea96c19118510c846"}, - {file = "msgpack-1.1.0-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:8cf9e8c3a2153934a23ac160cc4cba0ec035f6867c8013cc6077a79823370346"}, - {file = "msgpack-1.1.0-cp39-cp39-musllinux_1_2_i686.whl", hash = "sha256:3180065ec2abbe13a4ad37688b61b99d7f9e012a535b930e0e683ad6bc30155b"}, - {file = "msgpack-1.1.0-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:c5a91481a3cc573ac8c0d9aace09345d989dc4a0202b7fcb312c88c26d4e71a8"}, - {file = "msgpack-1.1.0-cp39-cp39-win32.whl", hash = "sha256:f80bc7d47f76089633763f952e67f8214cb7b3ee6bfa489b3cb6a84cfac114cd"}, - {file = "msgpack-1.1.0-cp39-cp39-win_amd64.whl", hash = "sha256:4d1b7ff2d6146e16e8bd665ac726a89c74163ef8cd39fa8c1087d4e52d3a2325"}, - {file = "msgpack-1.1.0.tar.gz", hash = "sha256:dd432ccc2c72b914e4cb77afce64aab761c1137cc698be3984eee260bcb2896e"}, + {file = "msgpack-1.1.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:353b6fc0c36fde68b661a12949d7d49f8f51ff5fa019c1e47c87c4ff34b080ed"}, + {file = "msgpack-1.1.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:79c408fcf76a958491b4e3b103d1c417044544b68e96d06432a189b43d1215c8"}, + {file = "msgpack-1.1.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:78426096939c2c7482bf31ef15ca219a9e24460289c00dd0b94411040bb73ad2"}, + {file = "msgpack-1.1.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8b17ba27727a36cb73aabacaa44b13090feb88a01d012c0f4be70c00f75048b4"}, + {file = "msgpack-1.1.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:7a17ac1ea6ec3c7687d70201cfda3b1e8061466f28f686c24f627cae4ea8efd0"}, + {file = "msgpack-1.1.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:88d1e966c9235c1d4e2afac21ca83933ba59537e2e2727a999bf3f515ca2af26"}, + {file = "msgpack-1.1.1-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:f6d58656842e1b2ddbe07f43f56b10a60f2ba5826164910968f5933e5178af75"}, + {file = "msgpack-1.1.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:96decdfc4adcbc087f5ea7ebdcfd3dee9a13358cae6e81d54be962efc38f6338"}, + {file = "msgpack-1.1.1-cp310-cp310-win32.whl", hash = "sha256:6640fd979ca9a212e4bcdf6eb74051ade2c690b862b679bfcb60ae46e6dc4bfd"}, + {file = "msgpack-1.1.1-cp310-cp310-win_amd64.whl", hash = "sha256:8b65b53204fe1bd037c40c4148d00ef918eb2108d24c9aaa20bc31f9810ce0a8"}, + {file = "msgpack-1.1.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:71ef05c1726884e44f8b1d1773604ab5d4d17729d8491403a705e649116c9558"}, + {file = "msgpack-1.1.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:36043272c6aede309d29d56851f8841ba907a1a3d04435e43e8a19928e243c1d"}, + {file = "msgpack-1.1.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a32747b1b39c3ac27d0670122b57e6e57f28eefb725e0b625618d1b59bf9d1e0"}, + {file = "msgpack-1.1.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8a8b10fdb84a43e50d38057b06901ec9da52baac6983d3f709d8507f3889d43f"}, + {file = "msgpack-1.1.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ba0c325c3f485dc54ec298d8b024e134acf07c10d494ffa24373bea729acf704"}, + {file = "msgpack-1.1.1-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:88daaf7d146e48ec71212ce21109b66e06a98e5e44dca47d853cbfe171d6c8d2"}, + {file = "msgpack-1.1.1-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:d8b55ea20dc59b181d3f47103f113e6f28a5e1c89fd5b67b9140edb442ab67f2"}, + {file = "msgpack-1.1.1-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:4a28e8072ae9779f20427af07f53bbb8b4aa81151054e882aee333b158da8752"}, + {file = "msgpack-1.1.1-cp311-cp311-win32.whl", hash = "sha256:7da8831f9a0fdb526621ba09a281fadc58ea12701bc709e7b8cbc362feabc295"}, + {file = "msgpack-1.1.1-cp311-cp311-win_amd64.whl", hash = "sha256:5fd1b58e1431008a57247d6e7cc4faa41c3607e8e7d4aaf81f7c29ea013cb458"}, + {file = "msgpack-1.1.1-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:ae497b11f4c21558d95de9f64fff7053544f4d1a17731c866143ed6bb4591238"}, + {file = "msgpack-1.1.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:33be9ab121df9b6b461ff91baac6f2731f83d9b27ed948c5b9d1978ae28bf157"}, + {file = "msgpack-1.1.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:6f64ae8fe7ffba251fecb8408540c34ee9df1c26674c50c4544d72dbf792e5ce"}, + {file = "msgpack-1.1.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a494554874691720ba5891c9b0b39474ba43ffb1aaf32a5dac874effb1619e1a"}, + {file = "msgpack-1.1.1-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:cb643284ab0ed26f6957d969fe0dd8bb17beb567beb8998140b5e38a90974f6c"}, + {file = "msgpack-1.1.1-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:d275a9e3c81b1093c060c3837e580c37f47c51eca031f7b5fb76f7b8470f5f9b"}, + {file = "msgpack-1.1.1-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:4fd6b577e4541676e0cc9ddc1709d25014d3ad9a66caa19962c4f5de30fc09ef"}, + {file = "msgpack-1.1.1-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:bb29aaa613c0a1c40d1af111abf025f1732cab333f96f285d6a93b934738a68a"}, + {file = "msgpack-1.1.1-cp312-cp312-win32.whl", hash = "sha256:870b9a626280c86cff9c576ec0d9cbcc54a1e5ebda9cd26dab12baf41fee218c"}, + {file = "msgpack-1.1.1-cp312-cp312-win_amd64.whl", hash = "sha256:5692095123007180dca3e788bb4c399cc26626da51629a31d40207cb262e67f4"}, + {file = "msgpack-1.1.1-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:3765afa6bd4832fc11c3749be4ba4b69a0e8d7b728f78e68120a157a4c5d41f0"}, + {file = "msgpack-1.1.1-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:8ddb2bcfd1a8b9e431c8d6f4f7db0773084e107730ecf3472f1dfe9ad583f3d9"}, + {file = "msgpack-1.1.1-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:196a736f0526a03653d829d7d4c5500a97eea3648aebfd4b6743875f28aa2af8"}, + {file = "msgpack-1.1.1-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9d592d06e3cc2f537ceeeb23d38799c6ad83255289bb84c2e5792e5a8dea268a"}, + {file = "msgpack-1.1.1-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4df2311b0ce24f06ba253fda361f938dfecd7b961576f9be3f3fbd60e87130ac"}, + {file = "msgpack-1.1.1-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:e4141c5a32b5e37905b5940aacbc59739f036930367d7acce7a64e4dec1f5e0b"}, + {file = "msgpack-1.1.1-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:b1ce7f41670c5a69e1389420436f41385b1aa2504c3b0c30620764b15dded2e7"}, + {file = "msgpack-1.1.1-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:4147151acabb9caed4e474c3344181e91ff7a388b888f1e19ea04f7e73dc7ad5"}, + {file = "msgpack-1.1.1-cp313-cp313-win32.whl", hash = "sha256:500e85823a27d6d9bba1d057c871b4210c1dd6fb01fbb764e37e4e8847376323"}, + {file = "msgpack-1.1.1-cp313-cp313-win_amd64.whl", hash = "sha256:6d489fba546295983abd142812bda76b57e33d0b9f5d5b71c09a583285506f69"}, + {file = "msgpack-1.1.1-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bba1be28247e68994355e028dcd668316db30c1f758d3241a7b903ac78dcd285"}, + {file = "msgpack-1.1.1-cp38-cp38-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b8f93dcddb243159c9e4109c9750ba5b335ab8d48d9522c5308cd05d7e3ce600"}, + {file = "msgpack-1.1.1-cp38-cp38-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:2fbbc0b906a24038c9958a1ba7ae0918ad35b06cb449d398b76a7d08470b0ed9"}, + {file = "msgpack-1.1.1-cp38-cp38-musllinux_1_2_aarch64.whl", hash = "sha256:61e35a55a546a1690d9d09effaa436c25ae6130573b6ee9829c37ef0f18d5e78"}, + {file = "msgpack-1.1.1-cp38-cp38-musllinux_1_2_i686.whl", hash = "sha256:1abfc6e949b352dadf4bce0eb78023212ec5ac42f6abfd469ce91d783c149c2a"}, + {file = "msgpack-1.1.1-cp38-cp38-musllinux_1_2_x86_64.whl", hash = "sha256:996f2609ddf0142daba4cefd767d6db26958aac8439ee41db9cc0db9f4c4c3a6"}, + {file = "msgpack-1.1.1-cp38-cp38-win32.whl", hash = "sha256:4d3237b224b930d58e9d83c81c0dba7aacc20fcc2f89c1e5423aa0529a4cd142"}, + {file = "msgpack-1.1.1-cp38-cp38-win_amd64.whl", hash = "sha256:da8f41e602574ece93dbbda1fab24650d6bf2a24089f9e9dbb4f5730ec1e58ad"}, + {file = "msgpack-1.1.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:f5be6b6bc52fad84d010cb45433720327ce886009d862f46b26d4d154001994b"}, + {file = "msgpack-1.1.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:3a89cd8c087ea67e64844287ea52888239cbd2940884eafd2dcd25754fb72232"}, + {file = "msgpack-1.1.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1d75f3807a9900a7d575d8d6674a3a47e9f227e8716256f35bc6f03fc597ffbf"}, + {file = "msgpack-1.1.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d182dac0221eb8faef2e6f44701812b467c02674a322c739355c39e94730cdbf"}, + {file = "msgpack-1.1.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:1b13fe0fb4aac1aa5320cd693b297fe6fdef0e7bea5518cbc2dd5299f873ae90"}, + {file = "msgpack-1.1.1-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:435807eeb1bc791ceb3247d13c79868deb22184e1fc4224808750f0d7d1affc1"}, + {file = "msgpack-1.1.1-cp39-cp39-musllinux_1_2_i686.whl", hash = "sha256:4835d17af722609a45e16037bb1d4d78b7bdf19d6c0128116d178956618c4e88"}, + {file = "msgpack-1.1.1-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:a8ef6e342c137888ebbfb233e02b8fbd689bb5b5fcc59b34711ac47ebd504478"}, + {file = "msgpack-1.1.1-cp39-cp39-win32.whl", hash = "sha256:61abccf9de335d9efd149e2fff97ed5974f2481b3353772e8e2dd3402ba2bd57"}, + {file = "msgpack-1.1.1-cp39-cp39-win_amd64.whl", hash = "sha256:40eae974c873b2992fd36424a5d9407f93e97656d999f43fca9d29f820899084"}, + {file = "msgpack-1.1.1.tar.gz", hash = "sha256:77b79ce34a2bdab2594f490c8e80dd62a02d650b91a75159a63ec413b8d104cd"}, ] [[package]] name = "multidict" -version = "6.4.4" +version = "6.6.3" description = "multidict implementation" optional = false python-versions = ">=3.9" groups = ["main"] files = [ - {file = "multidict-6.4.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:8adee3ac041145ffe4488ea73fa0a622b464cc25340d98be76924d0cda8545ff"}, - {file = "multidict-6.4.4-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:b61e98c3e2a861035aaccd207da585bdcacef65fe01d7a0d07478efac005e028"}, - {file = "multidict-6.4.4-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:75493f28dbadecdbb59130e74fe935288813301a8554dc32f0c631b6bdcdf8b0"}, - {file = "multidict-6.4.4-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4ffc3c6a37e048b5395ee235e4a2a0d639c2349dffa32d9367a42fc20d399772"}, - {file = "multidict-6.4.4-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:87cb72263946b301570b0f63855569a24ee8758aaae2cd182aae7d95fbc92ca7"}, - {file = "multidict-6.4.4-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9bbf7bd39822fd07e3609b6b4467af4c404dd2b88ee314837ad1830a7f4a8299"}, - {file = "multidict-6.4.4-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:d1f7cbd4f1f44ddf5fd86a8675b7679176eae770f2fc88115d6dddb6cefb59bc"}, - {file = "multidict-6.4.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:bb5ac9e5bfce0e6282e7f59ff7b7b9a74aa8e5c60d38186a4637f5aa764046ad"}, - {file = "multidict-6.4.4-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:4efc31dfef8c4eeb95b6b17d799eedad88c4902daba39ce637e23a17ea078915"}, - {file = "multidict-6.4.4-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:9fcad2945b1b91c29ef2b4050f590bfcb68d8ac8e0995a74e659aa57e8d78e01"}, - {file = "multidict-6.4.4-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:d877447e7368c7320832acb7159557e49b21ea10ffeb135c1077dbbc0816b598"}, - {file = "multidict-6.4.4-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:33a12ebac9f380714c298cbfd3e5b9c0c4e89c75fe612ae496512ee51028915f"}, - {file = "multidict-6.4.4-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:0f14ea68d29b43a9bf37953881b1e3eb75b2739e896ba4a6aa4ad4c5b9ffa145"}, - {file = "multidict-6.4.4-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:0327ad2c747a6600e4797d115d3c38a220fdb28e54983abe8964fd17e95ae83c"}, - {file = "multidict-6.4.4-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:d1a20707492db9719a05fc62ee215fd2c29b22b47c1b1ba347f9abc831e26683"}, - {file = "multidict-6.4.4-cp310-cp310-win32.whl", hash = "sha256:d83f18315b9fca5db2452d1881ef20f79593c4aa824095b62cb280019ef7aa3d"}, - {file = "multidict-6.4.4-cp310-cp310-win_amd64.whl", hash = "sha256:9c17341ee04545fd962ae07330cb5a39977294c883485c8d74634669b1f7fe04"}, - {file = "multidict-6.4.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:4f5f29794ac0e73d2a06ac03fd18870adc0135a9d384f4a306a951188ed02f95"}, - {file = "multidict-6.4.4-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:c04157266344158ebd57b7120d9b0b35812285d26d0e78193e17ef57bfe2979a"}, - {file = "multidict-6.4.4-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bb61ffd3ab8310d93427e460f565322c44ef12769f51f77277b4abad7b6f7223"}, - {file = "multidict-6.4.4-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:5e0ba18a9afd495f17c351d08ebbc4284e9c9f7971d715f196b79636a4d0de44"}, - {file = "multidict-6.4.4-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:9faf1b1dcaadf9f900d23a0e6d6c8eadd6a95795a0e57fcca73acce0eb912065"}, - {file = "multidict-6.4.4-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:a4d1cb1327c6082c4fce4e2a438483390964c02213bc6b8d782cf782c9b1471f"}, - {file = "multidict-6.4.4-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:941f1bec2f5dbd51feeb40aea654c2747f811ab01bdd3422a48a4e4576b7d76a"}, - {file = "multidict-6.4.4-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:e5f8a146184da7ea12910a4cec51ef85e44f6268467fb489c3caf0cd512f29c2"}, - {file = "multidict-6.4.4-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:232b7237e57ec3c09be97206bfb83a0aa1c5d7d377faa019c68a210fa35831f1"}, - {file = "multidict-6.4.4-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:55ae0721c1513e5e3210bca4fc98456b980b0c2c016679d3d723119b6b202c42"}, - {file = "multidict-6.4.4-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:51d662c072579f63137919d7bb8fc250655ce79f00c82ecf11cab678f335062e"}, - {file = "multidict-6.4.4-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:0e05c39962baa0bb19a6b210e9b1422c35c093b651d64246b6c2e1a7e242d9fd"}, - {file = "multidict-6.4.4-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:d5b1cc3ab8c31d9ebf0faa6e3540fb91257590da330ffe6d2393d4208e638925"}, - {file = "multidict-6.4.4-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:93ec84488a384cd7b8a29c2c7f467137d8a73f6fe38bb810ecf29d1ade011a7c"}, - {file = "multidict-6.4.4-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:b308402608493638763abc95f9dc0030bbd6ac6aff784512e8ac3da73a88af08"}, - {file = "multidict-6.4.4-cp311-cp311-win32.whl", hash = "sha256:343892a27d1a04d6ae455ecece12904d242d299ada01633d94c4f431d68a8c49"}, - {file = "multidict-6.4.4-cp311-cp311-win_amd64.whl", hash = "sha256:73484a94f55359780c0f458bbd3c39cb9cf9c182552177d2136e828269dee529"}, - {file = "multidict-6.4.4-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:dc388f75a1c00000824bf28b7633e40854f4127ede80512b44c3cfeeea1839a2"}, - {file = "multidict-6.4.4-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:98af87593a666f739d9dba5d0ae86e01b0e1a9cfcd2e30d2d361fbbbd1a9162d"}, - {file = "multidict-6.4.4-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:aff4cafea2d120327d55eadd6b7f1136a8e5a0ecf6fb3b6863e8aca32cd8e50a"}, - {file = "multidict-6.4.4-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:169c4ba7858176b797fe551d6e99040c531c775d2d57b31bcf4de6d7a669847f"}, - {file = "multidict-6.4.4-cp312-cp312-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:b9eb4c59c54421a32b3273d4239865cb14ead53a606db066d7130ac80cc8ec93"}, - {file = "multidict-6.4.4-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:7cf3bd54c56aa16fdb40028d545eaa8d051402b61533c21e84046e05513d5780"}, - {file = "multidict-6.4.4-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:f682c42003c7264134bfe886376299db4cc0c6cd06a3295b41b347044bcb5482"}, - {file = "multidict-6.4.4-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a920f9cf2abdf6e493c519492d892c362007f113c94da4c239ae88429835bad1"}, - {file = "multidict-6.4.4-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:530d86827a2df6504526106b4c104ba19044594f8722d3e87714e847c74a0275"}, - {file = "multidict-6.4.4-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:ecde56ea2439b96ed8a8d826b50c57364612ddac0438c39e473fafad7ae1c23b"}, - {file = "multidict-6.4.4-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:dc8c9736d8574b560634775ac0def6bdc1661fc63fa27ffdfc7264c565bcb4f2"}, - {file = "multidict-6.4.4-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:7f3d3b3c34867579ea47cbd6c1f2ce23fbfd20a273b6f9e3177e256584f1eacc"}, - {file = "multidict-6.4.4-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:87a728af265e08f96b6318ebe3c0f68b9335131f461efab2fc64cc84a44aa6ed"}, - {file = "multidict-6.4.4-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:9f193eeda1857f8e8d3079a4abd258f42ef4a4bc87388452ed1e1c4d2b0c8740"}, - {file = "multidict-6.4.4-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:be06e73c06415199200e9a2324a11252a3d62030319919cde5e6950ffeccf72e"}, - {file = "multidict-6.4.4-cp312-cp312-win32.whl", hash = "sha256:622f26ea6a7e19b7c48dd9228071f571b2fbbd57a8cd71c061e848f281550e6b"}, - {file = "multidict-6.4.4-cp312-cp312-win_amd64.whl", hash = "sha256:5e2bcda30d5009996ff439e02a9f2b5c3d64a20151d34898c000a6281faa3781"}, - {file = "multidict-6.4.4-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:82ffabefc8d84c2742ad19c37f02cde5ec2a1ee172d19944d380f920a340e4b9"}, - {file = "multidict-6.4.4-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:6a2f58a66fe2c22615ad26156354005391e26a2f3721c3621504cd87c1ea87bf"}, - {file = "multidict-6.4.4-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:5883d6ee0fd9d8a48e9174df47540b7545909841ac82354c7ae4cbe9952603bd"}, - {file = "multidict-6.4.4-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9abcf56a9511653fa1d052bfc55fbe53dbee8f34e68bd6a5a038731b0ca42d15"}, - {file = "multidict-6.4.4-cp313-cp313-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:6ed5ae5605d4ad5a049fad2a28bb7193400700ce2f4ae484ab702d1e3749c3f9"}, - {file = "multidict-6.4.4-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:bbfcb60396f9bcfa63e017a180c3105b8c123a63e9d1428a36544e7d37ca9e20"}, - {file = "multidict-6.4.4-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:b0f1987787f5f1e2076b59692352ab29a955b09ccc433c1f6b8e8e18666f608b"}, - {file = "multidict-6.4.4-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1d0121ccce8c812047d8d43d691a1ad7641f72c4f730474878a5aeae1b8ead8c"}, - {file = "multidict-6.4.4-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:83ec4967114295b8afd120a8eec579920c882831a3e4c3331d591a8e5bfbbc0f"}, - {file = "multidict-6.4.4-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:995f985e2e268deaf17867801b859a282e0448633f1310e3704b30616d269d69"}, - {file = "multidict-6.4.4-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:d832c608f94b9f92a0ec8b7e949be7792a642b6e535fcf32f3e28fab69eeb046"}, - {file = "multidict-6.4.4-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:d21c1212171cf7da703c5b0b7a0e85be23b720818aef502ad187d627316d5645"}, - {file = "multidict-6.4.4-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:cbebaa076aaecad3d4bb4c008ecc73b09274c952cf6a1b78ccfd689e51f5a5b0"}, - {file = "multidict-6.4.4-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:c93a6fb06cc8e5d3628b2b5fda215a5db01e8f08fc15fadd65662d9b857acbe4"}, - {file = "multidict-6.4.4-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:8cd8f81f1310182362fb0c7898145ea9c9b08a71081c5963b40ee3e3cac589b1"}, - {file = "multidict-6.4.4-cp313-cp313-win32.whl", hash = "sha256:3e9f1cd61a0ab857154205fb0b1f3d3ace88d27ebd1409ab7af5096e409614cd"}, - {file = "multidict-6.4.4-cp313-cp313-win_amd64.whl", hash = "sha256:8ffb40b74400e4455785c2fa37eba434269149ec525fc8329858c862e4b35373"}, - {file = "multidict-6.4.4-cp313-cp313t-macosx_10_13_universal2.whl", hash = "sha256:6a602151dbf177be2450ef38966f4be3467d41a86c6a845070d12e17c858a156"}, - {file = "multidict-6.4.4-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:0d2b9712211b860d123815a80b859075d86a4d54787e247d7fbee9db6832cf1c"}, - {file = "multidict-6.4.4-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:d2fa86af59f8fc1972e121ade052145f6da22758f6996a197d69bb52f8204e7e"}, - {file = "multidict-6.4.4-cp313-cp313t-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:50855d03e9e4d66eab6947ba688ffb714616f985838077bc4b490e769e48da51"}, - {file = "multidict-6.4.4-cp313-cp313t-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:5bce06b83be23225be1905dcdb6b789064fae92499fbc458f59a8c0e68718601"}, - {file = "multidict-6.4.4-cp313-cp313t-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:66ed0731f8e5dfd8369a883b6e564aca085fb9289aacabd9decd70568b9a30de"}, - {file = "multidict-6.4.4-cp313-cp313t-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:329ae97fc2f56f44d91bc47fe0972b1f52d21c4b7a2ac97040da02577e2daca2"}, - {file = "multidict-6.4.4-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c27e5dcf520923d6474d98b96749e6805f7677e93aaaf62656005b8643f907ab"}, - {file = "multidict-6.4.4-cp313-cp313t-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:058cc59b9e9b143cc56715e59e22941a5d868c322242278d28123a5d09cdf6b0"}, - {file = "multidict-6.4.4-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:69133376bc9a03f8c47343d33f91f74a99c339e8b58cea90433d8e24bb298031"}, - {file = "multidict-6.4.4-cp313-cp313t-musllinux_1_2_armv7l.whl", hash = "sha256:d6b15c55721b1b115c5ba178c77104123745b1417527ad9641a4c5e2047450f0"}, - {file = "multidict-6.4.4-cp313-cp313t-musllinux_1_2_i686.whl", hash = "sha256:a887b77f51d3d41e6e1a63cf3bc7ddf24de5939d9ff69441387dfefa58ac2e26"}, - {file = "multidict-6.4.4-cp313-cp313t-musllinux_1_2_ppc64le.whl", hash = "sha256:632a3bf8f1787f7ef7d3c2f68a7bde5be2f702906f8b5842ad6da9d974d0aab3"}, - {file = "multidict-6.4.4-cp313-cp313t-musllinux_1_2_s390x.whl", hash = "sha256:a145c550900deb7540973c5cdb183b0d24bed6b80bf7bddf33ed8f569082535e"}, - {file = "multidict-6.4.4-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:cc5d83c6619ca5c9672cb78b39ed8542f1975a803dee2cda114ff73cbb076edd"}, - {file = "multidict-6.4.4-cp313-cp313t-win32.whl", hash = "sha256:3312f63261b9df49be9d57aaa6abf53a6ad96d93b24f9cc16cf979956355ce6e"}, - {file = "multidict-6.4.4-cp313-cp313t-win_amd64.whl", hash = "sha256:ba852168d814b2c73333073e1c7116d9395bea69575a01b0b3c89d2d5a87c8fb"}, - {file = "multidict-6.4.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:603f39bd1cf85705c6c1ba59644b480dfe495e6ee2b877908de93322705ad7cf"}, - {file = "multidict-6.4.4-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:fc60f91c02e11dfbe3ff4e1219c085695c339af72d1641800fe6075b91850c8f"}, - {file = "multidict-6.4.4-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:496bcf01c76a70a31c3d746fd39383aad8d685ce6331e4c709e9af4ced5fa221"}, - {file = "multidict-6.4.4-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4219390fb5bf8e548e77b428bb36a21d9382960db5321b74d9d9987148074d6b"}, - {file = "multidict-6.4.4-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:3ef4e9096ff86dfdcbd4a78253090ba13b1d183daa11b973e842465d94ae1772"}, - {file = "multidict-6.4.4-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:49a29d7133b1fc214e818bbe025a77cc6025ed9a4f407d2850373ddde07fd04a"}, - {file = "multidict-6.4.4-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e32053d6d3a8b0dfe49fde05b496731a0e6099a4df92154641c00aa76786aef5"}, - {file = "multidict-6.4.4-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8cc403092a49509e8ef2d2fd636a8ecefc4698cc57bbe894606b14579bc2a955"}, - {file = "multidict-6.4.4-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:5363f9b2a7f3910e5c87d8b1855c478c05a2dc559ac57308117424dfaad6805c"}, - {file = "multidict-6.4.4-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:2e543a40e4946cf70a88a3be87837a3ae0aebd9058ba49e91cacb0b2cd631e2b"}, - {file = "multidict-6.4.4-cp39-cp39-musllinux_1_2_armv7l.whl", hash = "sha256:60d849912350da557fe7de20aa8cf394aada6980d0052cc829eeda4a0db1c1db"}, - {file = "multidict-6.4.4-cp39-cp39-musllinux_1_2_i686.whl", hash = "sha256:19d08b4f22eae45bb018b9f06e2838c1e4b853c67628ef8ae126d99de0da6395"}, - {file = "multidict-6.4.4-cp39-cp39-musllinux_1_2_ppc64le.whl", hash = "sha256:d693307856d1ef08041e8b6ff01d5b4618715007d288490ce2c7e29013c12b9a"}, - {file = "multidict-6.4.4-cp39-cp39-musllinux_1_2_s390x.whl", hash = "sha256:fad6daaed41021934917f4fb03ca2db8d8a4d79bf89b17ebe77228eb6710c003"}, - {file = "multidict-6.4.4-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:c10d17371bff801af0daf8b073c30b6cf14215784dc08cd5c43ab5b7b8029bbc"}, - {file = "multidict-6.4.4-cp39-cp39-win32.whl", hash = "sha256:7e23f2f841fcb3ebd4724a40032d32e0892fbba4143e43d2a9e7695c5e50e6bd"}, - {file = "multidict-6.4.4-cp39-cp39-win_amd64.whl", hash = "sha256:4d7b50b673ffb4ff4366e7ab43cf1f0aef4bd3608735c5fbdf0bdb6f690da411"}, - {file = "multidict-6.4.4-py3-none-any.whl", hash = "sha256:bd4557071b561a8b3b6075c3ce93cf9bfb6182cb241805c3d66ced3b75eff4ac"}, - {file = "multidict-6.4.4.tar.gz", hash = "sha256:69ee9e6ba214b5245031b76233dd95408a0fd57fdb019ddcc1ead4790932a8e8"}, + {file = "multidict-6.6.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:a2be5b7b35271f7fff1397204ba6708365e3d773579fe2a30625e16c4b4ce817"}, + {file = "multidict-6.6.3-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:12f4581d2930840295c461764b9a65732ec01250b46c6b2c510d7ee68872b140"}, + {file = "multidict-6.6.3-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:dd7793bab517e706c9ed9d7310b06c8672fd0aeee5781bfad612f56b8e0f7d14"}, + {file = "multidict-6.6.3-cp310-cp310-manylinux1_i686.manylinux2014_i686.manylinux_2_17_i686.manylinux_2_5_i686.whl", hash = "sha256:72d8815f2cd3cf3df0f83cac3f3ef801d908b2d90409ae28102e0553af85545a"}, + {file = "multidict-6.6.3-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:531e331a2ee53543ab32b16334e2deb26f4e6b9b28e41f8e0c87e99a6c8e2d69"}, + {file = "multidict-6.6.3-cp310-cp310-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:42ca5aa9329a63be8dc49040f63817d1ac980e02eeddba763a9ae5b4027b9c9c"}, + {file = "multidict-6.6.3-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:208b9b9757060b9faa6f11ab4bc52846e4f3c2fb8b14d5680c8aac80af3dc751"}, + {file = "multidict-6.6.3-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:acf6b97bd0884891af6a8b43d0f586ab2fcf8e717cbd47ab4bdddc09e20652d8"}, + {file = "multidict-6.6.3-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:68e9e12ed00e2089725669bdc88602b0b6f8d23c0c95e52b95f0bc69f7fe9b55"}, + {file = "multidict-6.6.3-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:05db2f66c9addb10cfa226e1acb363450fab2ff8a6df73c622fefe2f5af6d4e7"}, + {file = "multidict-6.6.3-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:0db58da8eafb514db832a1b44f8fa7906fdd102f7d982025f816a93ba45e3dcb"}, + {file = "multidict-6.6.3-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:14117a41c8fdb3ee19c743b1c027da0736fdb79584d61a766da53d399b71176c"}, + {file = "multidict-6.6.3-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:877443eaaabcd0b74ff32ebeed6f6176c71850feb7d6a1d2db65945256ea535c"}, + {file = "multidict-6.6.3-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:70b72e749a4f6e7ed8fb334fa8d8496384840319512746a5f42fa0aec79f4d61"}, + {file = "multidict-6.6.3-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:43571f785b86afd02b3855c5ac8e86ec921b760298d6f82ff2a61daf5a35330b"}, + {file = "multidict-6.6.3-cp310-cp310-win32.whl", hash = "sha256:20c5a0c3c13a15fd5ea86c42311859f970070e4e24de5a550e99d7c271d76318"}, + {file = "multidict-6.6.3-cp310-cp310-win_amd64.whl", hash = "sha256:ab0a34a007704c625e25a9116c6770b4d3617a071c8a7c30cd338dfbadfe6485"}, + {file = "multidict-6.6.3-cp310-cp310-win_arm64.whl", hash = "sha256:769841d70ca8bdd140a715746199fc6473414bd02efd678d75681d2d6a8986c5"}, + {file = "multidict-6.6.3-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:18f4eba0cbac3546b8ae31e0bbc55b02c801ae3cbaf80c247fcdd89b456ff58c"}, + {file = "multidict-6.6.3-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:ef43b5dd842382329e4797c46f10748d8c2b6e0614f46b4afe4aee9ac33159df"}, + {file = "multidict-6.6.3-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:bf9bd1fd5eec01494e0f2e8e446a74a85d5e49afb63d75a9934e4a5423dba21d"}, + {file = "multidict-6.6.3-cp311-cp311-manylinux1_i686.manylinux2014_i686.manylinux_2_17_i686.manylinux_2_5_i686.whl", hash = "sha256:5bd8d6f793a787153956cd35e24f60485bf0651c238e207b9a54f7458b16d539"}, + {file = "multidict-6.6.3-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:1bf99b4daf908c73856bd87ee0a2499c3c9a3d19bb04b9c6025e66af3fd07462"}, + {file = "multidict-6.6.3-cp311-cp311-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:0b9e59946b49dafaf990fd9c17ceafa62976e8471a14952163d10a7a630413a9"}, + {file = "multidict-6.6.3-cp311-cp311-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:e2db616467070d0533832d204c54eea6836a5e628f2cb1e6dfd8cd6ba7277cb7"}, + {file = "multidict-6.6.3-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:7394888236621f61dcdd25189b2768ae5cc280f041029a5bcf1122ac63df79f9"}, + {file = "multidict-6.6.3-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:f114d8478733ca7388e7c7e0ab34b72547476b97009d643644ac33d4d3fe1821"}, + {file = "multidict-6.6.3-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:cdf22e4db76d323bcdc733514bf732e9fb349707c98d341d40ebcc6e9318ef3d"}, + {file = "multidict-6.6.3-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:e995a34c3d44ab511bfc11aa26869b9d66c2d8c799fa0e74b28a473a692532d6"}, + {file = "multidict-6.6.3-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:766a4a5996f54361d8d5a9050140aa5362fe48ce51c755a50c0bc3706460c430"}, + {file = "multidict-6.6.3-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:3893a0d7d28a7fe6ca7a1f760593bc13038d1d35daf52199d431b61d2660602b"}, + {file = "multidict-6.6.3-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:934796c81ea996e61914ba58064920d6cad5d99140ac3167901eb932150e2e56"}, + {file = "multidict-6.6.3-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:9ed948328aec2072bc00f05d961ceadfd3e9bfc2966c1319aeaf7b7c21219183"}, + {file = "multidict-6.6.3-cp311-cp311-win32.whl", hash = "sha256:9f5b28c074c76afc3e4c610c488e3493976fe0e596dd3db6c8ddfbb0134dcac5"}, + {file = "multidict-6.6.3-cp311-cp311-win_amd64.whl", hash = "sha256:bc7f6fbc61b1c16050a389c630da0b32fc6d4a3d191394ab78972bf5edc568c2"}, + {file = "multidict-6.6.3-cp311-cp311-win_arm64.whl", hash = "sha256:d4e47d8faffaae822fb5cba20937c048d4f734f43572e7079298a6c39fb172cb"}, + {file = "multidict-6.6.3-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:056bebbeda16b2e38642d75e9e5310c484b7c24e3841dc0fb943206a72ec89d6"}, + {file = "multidict-6.6.3-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:e5f481cccb3c5c5e5de5d00b5141dc589c1047e60d07e85bbd7dea3d4580d63f"}, + {file = "multidict-6.6.3-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:10bea2ee839a759ee368b5a6e47787f399b41e70cf0c20d90dfaf4158dfb4e55"}, + {file = "multidict-6.6.3-cp312-cp312-manylinux1_i686.manylinux2014_i686.manylinux_2_17_i686.manylinux_2_5_i686.whl", hash = "sha256:2334cfb0fa9549d6ce2c21af2bfbcd3ac4ec3646b1b1581c88e3e2b1779ec92b"}, + {file = "multidict-6.6.3-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b8fee016722550a2276ca2cb5bb624480e0ed2bd49125b2b73b7010b9090e888"}, + {file = "multidict-6.6.3-cp312-cp312-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:e5511cb35f5c50a2db21047c875eb42f308c5583edf96bd8ebf7d770a9d68f6d"}, + {file = "multidict-6.6.3-cp312-cp312-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:712b348f7f449948e0a6c4564a21c7db965af900973a67db432d724619b3c680"}, + {file = "multidict-6.6.3-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:e4e15d2138ee2694e038e33b7c3da70e6b0ad8868b9f8094a72e1414aeda9c1a"}, + {file = "multidict-6.6.3-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:8df25594989aebff8a130f7899fa03cbfcc5d2b5f4a461cf2518236fe6f15961"}, + {file = "multidict-6.6.3-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:159ca68bfd284a8860f8d8112cf0521113bffd9c17568579e4d13d1f1dc76b65"}, + {file = "multidict-6.6.3-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:e098c17856a8c9ade81b4810888c5ad1914099657226283cab3062c0540b0643"}, + {file = "multidict-6.6.3-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:67c92ed673049dec52d7ed39f8cf9ebbadf5032c774058b4406d18c8f8fe7063"}, + {file = "multidict-6.6.3-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:bd0578596e3a835ef451784053cfd327d607fc39ea1a14812139339a18a0dbc3"}, + {file = "multidict-6.6.3-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:346055630a2df2115cd23ae271910b4cae40f4e336773550dca4889b12916e75"}, + {file = "multidict-6.6.3-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:555ff55a359302b79de97e0468e9ee80637b0de1fce77721639f7cd9440b3a10"}, + {file = "multidict-6.6.3-cp312-cp312-win32.whl", hash = "sha256:73ab034fb8d58ff85c2bcbadc470efc3fafeea8affcf8722855fb94557f14cc5"}, + {file = "multidict-6.6.3-cp312-cp312-win_amd64.whl", hash = "sha256:04cbcce84f63b9af41bad04a54d4cc4e60e90c35b9e6ccb130be2d75b71f8c17"}, + {file = "multidict-6.6.3-cp312-cp312-win_arm64.whl", hash = "sha256:0f1130b896ecb52d2a1e615260f3ea2af55fa7dc3d7c3003ba0c3121a759b18b"}, + {file = "multidict-6.6.3-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:540d3c06d48507357a7d57721e5094b4f7093399a0106c211f33540fdc374d55"}, + {file = "multidict-6.6.3-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:9c19cea2a690f04247d43f366d03e4eb110a0dc4cd1bbeee4d445435428ed35b"}, + {file = "multidict-6.6.3-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:7af039820cfd00effec86bda5d8debef711a3e86a1d3772e85bea0f243a4bd65"}, + {file = "multidict-6.6.3-cp313-cp313-manylinux1_i686.manylinux2014_i686.manylinux_2_17_i686.manylinux_2_5_i686.whl", hash = "sha256:500b84f51654fdc3944e936f2922114349bf8fdcac77c3092b03449f0e5bc2b3"}, + {file = "multidict-6.6.3-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:f3fc723ab8a5c5ed6c50418e9bfcd8e6dceba6c271cee6728a10a4ed8561520c"}, + {file = "multidict-6.6.3-cp313-cp313-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:94c47ea3ade005b5976789baaed66d4de4480d0a0bf31cef6edaa41c1e7b56a6"}, + {file = "multidict-6.6.3-cp313-cp313-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:dbc7cf464cc6d67e83e136c9f55726da3a30176f020a36ead246eceed87f1cd8"}, + {file = "multidict-6.6.3-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:900eb9f9da25ada070f8ee4a23f884e0ee66fe4e1a38c3af644256a508ad81ca"}, + {file = "multidict-6.6.3-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:7c6df517cf177da5d47ab15407143a89cd1a23f8b335f3a28d57e8b0a3dbb884"}, + {file = "multidict-6.6.3-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:4ef421045f13879e21c994b36e728d8e7d126c91a64b9185810ab51d474f27e7"}, + {file = "multidict-6.6.3-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:6c1e61bb4f80895c081790b6b09fa49e13566df8fbff817da3f85b3a8192e36b"}, + {file = "multidict-6.6.3-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:e5e8523bb12d7623cd8300dbd91b9e439a46a028cd078ca695eb66ba31adee3c"}, + {file = "multidict-6.6.3-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:ef58340cc896219e4e653dade08fea5c55c6df41bcc68122e3be3e9d873d9a7b"}, + {file = "multidict-6.6.3-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:fc9dc435ec8699e7b602b94fe0cd4703e69273a01cbc34409af29e7820f777f1"}, + {file = "multidict-6.6.3-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:9e864486ef4ab07db5e9cb997bad2b681514158d6954dd1958dfb163b83d53e6"}, + {file = "multidict-6.6.3-cp313-cp313-win32.whl", hash = "sha256:5633a82fba8e841bc5c5c06b16e21529573cd654f67fd833650a215520a6210e"}, + {file = "multidict-6.6.3-cp313-cp313-win_amd64.whl", hash = "sha256:e93089c1570a4ad54c3714a12c2cef549dc9d58e97bcded193d928649cab78e9"}, + {file = "multidict-6.6.3-cp313-cp313-win_arm64.whl", hash = "sha256:c60b401f192e79caec61f166da9c924e9f8bc65548d4246842df91651e83d600"}, + {file = "multidict-6.6.3-cp313-cp313t-macosx_10_13_universal2.whl", hash = "sha256:02fd8f32d403a6ff13864b0851f1f523d4c988051eea0471d4f1fd8010f11134"}, + {file = "multidict-6.6.3-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:f3aa090106b1543f3f87b2041eef3c156c8da2aed90c63a2fbed62d875c49c37"}, + {file = "multidict-6.6.3-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:e924fb978615a5e33ff644cc42e6aa241effcf4f3322c09d4f8cebde95aff5f8"}, + {file = "multidict-6.6.3-cp313-cp313t-manylinux1_i686.manylinux2014_i686.manylinux_2_17_i686.manylinux_2_5_i686.whl", hash = "sha256:b9fe5a0e57c6dbd0e2ce81ca66272282c32cd11d31658ee9553849d91289e1c1"}, + {file = "multidict-6.6.3-cp313-cp313t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b24576f208793ebae00280c59927c3b7c2a3b1655e443a25f753c4611bc1c373"}, + {file = "multidict-6.6.3-cp313-cp313t-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:135631cb6c58eac37d7ac0df380294fecdc026b28837fa07c02e459c7fb9c54e"}, + {file = "multidict-6.6.3-cp313-cp313t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:274d416b0df887aef98f19f21578653982cfb8a05b4e187d4a17103322eeaf8f"}, + {file = "multidict-6.6.3-cp313-cp313t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:e252017a817fad7ce05cafbe5711ed40faeb580e63b16755a3a24e66fa1d87c0"}, + {file = "multidict-6.6.3-cp313-cp313t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:2e4cc8d848cd4fe1cdee28c13ea79ab0ed37fc2e89dd77bac86a2e7959a8c3bc"}, + {file = "multidict-6.6.3-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:9e236a7094b9c4c1b7585f6b9cca34b9d833cf079f7e4c49e6a4a6ec9bfdc68f"}, + {file = "multidict-6.6.3-cp313-cp313t-musllinux_1_2_armv7l.whl", hash = "sha256:e0cb0ab69915c55627c933f0b555a943d98ba71b4d1c57bc0d0a66e2567c7471"}, + {file = "multidict-6.6.3-cp313-cp313t-musllinux_1_2_i686.whl", hash = "sha256:81ef2f64593aba09c5212a3d0f8c906a0d38d710a011f2f42759704d4557d3f2"}, + {file = "multidict-6.6.3-cp313-cp313t-musllinux_1_2_ppc64le.whl", hash = "sha256:b9cbc60010de3562545fa198bfc6d3825df430ea96d2cc509c39bd71e2e7d648"}, + {file = "multidict-6.6.3-cp313-cp313t-musllinux_1_2_s390x.whl", hash = "sha256:70d974eaaa37211390cd02ef93b7e938de564bbffa866f0b08d07e5e65da783d"}, + {file = "multidict-6.6.3-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:3713303e4a6663c6d01d648a68f2848701001f3390a030edaaf3fc949c90bf7c"}, + {file = "multidict-6.6.3-cp313-cp313t-win32.whl", hash = "sha256:639ecc9fe7cd73f2495f62c213e964843826f44505a3e5d82805aa85cac6f89e"}, + {file = "multidict-6.6.3-cp313-cp313t-win_amd64.whl", hash = "sha256:9f97e181f344a0ef3881b573d31de8542cc0dbc559ec68c8f8b5ce2c2e91646d"}, + {file = "multidict-6.6.3-cp313-cp313t-win_arm64.whl", hash = "sha256:ce8b7693da41a3c4fde5871c738a81490cea5496c671d74374c8ab889e1834fb"}, + {file = "multidict-6.6.3-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:c8161b5a7778d3137ea2ee7ae8a08cce0010de3b00ac671c5ebddeaa17cefd22"}, + {file = "multidict-6.6.3-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:1328201ee930f069961ae707d59c6627ac92e351ed5b92397cf534d1336ce557"}, + {file = "multidict-6.6.3-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:b1db4d2093d6b235de76932febf9d50766cf49a5692277b2c28a501c9637f616"}, + {file = "multidict-6.6.3-cp39-cp39-manylinux1_i686.manylinux2014_i686.manylinux_2_17_i686.manylinux_2_5_i686.whl", hash = "sha256:53becb01dd8ebd19d1724bebe369cfa87e4e7f29abbbe5c14c98ce4c383e16cd"}, + {file = "multidict-6.6.3-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:41bb9d1d4c303886e2d85bade86e59885112a7f4277af5ad47ab919a2251f306"}, + {file = "multidict-6.6.3-cp39-cp39-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:775b464d31dac90f23192af9c291dc9f423101857e33e9ebf0020a10bfcf4144"}, + {file = "multidict-6.6.3-cp39-cp39-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:d04d01f0a913202205a598246cf77826fe3baa5a63e9f6ccf1ab0601cf56eca0"}, + {file = "multidict-6.6.3-cp39-cp39-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:d25594d3b38a2e6cabfdcafef339f754ca6e81fbbdb6650ad773ea9775af35ab"}, + {file = "multidict-6.6.3-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:35712f1748d409e0707b165bf49f9f17f9e28ae85470c41615778f8d4f7d9609"}, + {file = "multidict-6.6.3-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:1c8082e5814b662de8589d6a06c17e77940d5539080cbab9fe6794b5241b76d9"}, + {file = "multidict-6.6.3-cp39-cp39-musllinux_1_2_armv7l.whl", hash = "sha256:61af8a4b771f1d4d000b3168c12c3120ccf7284502a94aa58c68a81f5afac090"}, + {file = "multidict-6.6.3-cp39-cp39-musllinux_1_2_i686.whl", hash = "sha256:448e4a9afccbf297577f2eaa586f07067441e7b63c8362a3540ba5a38dc0f14a"}, + {file = "multidict-6.6.3-cp39-cp39-musllinux_1_2_ppc64le.whl", hash = "sha256:233ad16999afc2bbd3e534ad8dbe685ef8ee49a37dbc2cdc9514e57b6d589ced"}, + {file = "multidict-6.6.3-cp39-cp39-musllinux_1_2_s390x.whl", hash = "sha256:bb933c891cd4da6bdcc9733d048e994e22e1883287ff7540c2a0f3b117605092"}, + {file = "multidict-6.6.3-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:37b09ca60998e87734699e88c2363abfd457ed18cfbf88e4009a4e83788e63ed"}, + {file = "multidict-6.6.3-cp39-cp39-win32.whl", hash = "sha256:f54cb79d26d0cd420637d184af38f0668558f3c4bbe22ab7ad830e67249f2e0b"}, + {file = "multidict-6.6.3-cp39-cp39-win_amd64.whl", hash = "sha256:295adc9c0551e5d5214b45cf29ca23dbc28c2d197a9c30d51aed9e037cb7c578"}, + {file = "multidict-6.6.3-cp39-cp39-win_arm64.whl", hash = "sha256:15332783596f227db50fb261c2c251a58ac3873c457f3a550a95d5c0aa3c770d"}, + {file = "multidict-6.6.3-py3-none-any.whl", hash = "sha256:8db10f29c7541fc5da4defd8cd697e1ca429db743fa716325f236079b96f775a"}, + {file = "multidict-6.6.3.tar.gz", hash = "sha256:798a9eb12dab0a6c2e29c1de6f3468af5cb2da6053a20dfa3344907eed0937cc"}, ] [[package]] @@ -2333,75 +2368,75 @@ test = ["pytest", "pytest-console-scripts", "pytest-jupyter", "pytest-tornasync" [[package]] name = "numpy" -version = "2.3.0" +version = "2.3.1" description = "Fundamental package for array computing in Python" optional = false python-versions = ">=3.11" groups = ["main"] files = [ - {file = "numpy-2.3.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:c3c9fdde0fa18afa1099d6257eb82890ea4f3102847e692193b54e00312a9ae9"}, - {file = "numpy-2.3.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:46d16f72c2192da7b83984aa5455baee640e33a9f1e61e656f29adf55e406c2b"}, - {file = "numpy-2.3.0-cp311-cp311-macosx_14_0_arm64.whl", hash = "sha256:a0be278be9307c4ab06b788f2a077f05e180aea817b3e41cebbd5aaf7bd85ed3"}, - {file = "numpy-2.3.0-cp311-cp311-macosx_14_0_x86_64.whl", hash = "sha256:99224862d1412d2562248d4710126355d3a8db7672170a39d6909ac47687a8a4"}, - {file = "numpy-2.3.0-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:2393a914db64b0ead0ab80c962e42d09d5f385802006a6c87835acb1f58adb96"}, - {file = "numpy-2.3.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:7729c8008d55e80784bd113787ce876ca117185c579c0d626f59b87d433ea779"}, - {file = "numpy-2.3.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:06d4fb37a8d383b769281714897420c5cc3545c79dc427df57fc9b852ee0bf58"}, - {file = "numpy-2.3.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:c39ec392b5db5088259c68250e342612db82dc80ce044cf16496cf14cf6bc6f8"}, - {file = "numpy-2.3.0-cp311-cp311-win32.whl", hash = "sha256:ee9d3ee70d62827bc91f3ea5eee33153212c41f639918550ac0475e3588da59f"}, - {file = "numpy-2.3.0-cp311-cp311-win_amd64.whl", hash = "sha256:43c55b6a860b0eb44d42341438b03513cf3879cb3617afb749ad49307e164edd"}, - {file = "numpy-2.3.0-cp311-cp311-win_arm64.whl", hash = "sha256:2e6a1409eee0cb0316cb64640a49a49ca44deb1a537e6b1121dc7c458a1299a8"}, - {file = "numpy-2.3.0-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:389b85335838155a9076e9ad7f8fdba0827496ec2d2dc32ce69ce7898bde03ba"}, - {file = "numpy-2.3.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:9498f60cd6bb8238d8eaf468a3d5bb031d34cd12556af53510f05fcf581c1b7e"}, - {file = "numpy-2.3.0-cp312-cp312-macosx_14_0_arm64.whl", hash = "sha256:622a65d40d8eb427d8e722fd410ac3ad4958002f109230bc714fa551044ebae2"}, - {file = "numpy-2.3.0-cp312-cp312-macosx_14_0_x86_64.whl", hash = "sha256:b9446d9d8505aadadb686d51d838f2b6688c9e85636a0c3abaeb55ed54756459"}, - {file = "numpy-2.3.0-cp312-cp312-manylinux_2_28_aarch64.whl", hash = "sha256:50080245365d75137a2bf46151e975de63146ae6d79f7e6bd5c0e85c9931d06a"}, - {file = "numpy-2.3.0-cp312-cp312-manylinux_2_28_x86_64.whl", hash = "sha256:c24bb4113c66936eeaa0dc1e47c74770453d34f46ee07ae4efd853a2ed1ad10a"}, - {file = "numpy-2.3.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:4d8d294287fdf685281e671886c6dcdf0291a7c19db3e5cb4178d07ccf6ecc67"}, - {file = "numpy-2.3.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:6295f81f093b7f5769d1728a6bd8bf7466de2adfa771ede944ce6711382b89dc"}, - {file = "numpy-2.3.0-cp312-cp312-win32.whl", hash = "sha256:e6648078bdd974ef5d15cecc31b0c410e2e24178a6e10bf511e0557eed0f2570"}, - {file = "numpy-2.3.0-cp312-cp312-win_amd64.whl", hash = "sha256:0898c67a58cdaaf29994bc0e2c65230fd4de0ac40afaf1584ed0b02cd74c6fdd"}, - {file = "numpy-2.3.0-cp312-cp312-win_arm64.whl", hash = "sha256:bd8df082b6c4695753ad6193018c05aac465d634834dca47a3ae06d4bb22d9ea"}, - {file = "numpy-2.3.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:5754ab5595bfa2c2387d241296e0381c21f44a4b90a776c3c1d39eede13a746a"}, - {file = "numpy-2.3.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:d11fa02f77752d8099573d64e5fe33de3229b6632036ec08f7080f46b6649959"}, - {file = "numpy-2.3.0-cp313-cp313-macosx_14_0_arm64.whl", hash = "sha256:aba48d17e87688a765ab1cd557882052f238e2f36545dfa8e29e6a91aef77afe"}, - {file = "numpy-2.3.0-cp313-cp313-macosx_14_0_x86_64.whl", hash = "sha256:4dc58865623023b63b10d52f18abaac3729346a7a46a778381e0e3af4b7f3beb"}, - {file = "numpy-2.3.0-cp313-cp313-manylinux_2_28_aarch64.whl", hash = "sha256:df470d376f54e052c76517393fa443758fefcdd634645bc9c1f84eafc67087f0"}, - {file = "numpy-2.3.0-cp313-cp313-manylinux_2_28_x86_64.whl", hash = "sha256:87717eb24d4a8a64683b7a4e91ace04e2f5c7c77872f823f02a94feee186168f"}, - {file = "numpy-2.3.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:d8fa264d56882b59dcb5ea4d6ab6f31d0c58a57b41aec605848b6eb2ef4a43e8"}, - {file = "numpy-2.3.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:e651756066a0eaf900916497e20e02fe1ae544187cb0fe88de981671ee7f6270"}, - {file = "numpy-2.3.0-cp313-cp313-win32.whl", hash = "sha256:e43c3cce3b6ae5f94696669ff2a6eafd9a6b9332008bafa4117af70f4b88be6f"}, - {file = "numpy-2.3.0-cp313-cp313-win_amd64.whl", hash = "sha256:81ae0bf2564cf475f94be4a27ef7bcf8af0c3e28da46770fc904da9abd5279b5"}, - {file = "numpy-2.3.0-cp313-cp313-win_arm64.whl", hash = "sha256:c8738baa52505fa6e82778580b23f945e3578412554d937093eac9205e845e6e"}, - {file = "numpy-2.3.0-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:39b27d8b38942a647f048b675f134dd5a567f95bfff481f9109ec308515c51d8"}, - {file = "numpy-2.3.0-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:0eba4a1ea88f9a6f30f56fdafdeb8da3774349eacddab9581a21234b8535d3d3"}, - {file = "numpy-2.3.0-cp313-cp313t-macosx_14_0_arm64.whl", hash = "sha256:b0f1f11d0a1da54927436505a5a7670b154eac27f5672afc389661013dfe3d4f"}, - {file = "numpy-2.3.0-cp313-cp313t-macosx_14_0_x86_64.whl", hash = "sha256:690d0a5b60a47e1f9dcec7b77750a4854c0d690e9058b7bef3106e3ae9117808"}, - {file = "numpy-2.3.0-cp313-cp313t-manylinux_2_28_aarch64.whl", hash = "sha256:8b51ead2b258284458e570942137155978583e407babc22e3d0ed7af33ce06f8"}, - {file = "numpy-2.3.0-cp313-cp313t-manylinux_2_28_x86_64.whl", hash = "sha256:aaf81c7b82c73bd9b45e79cfb9476cb9c29e937494bfe9092c26aece812818ad"}, - {file = "numpy-2.3.0-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:f420033a20b4f6a2a11f585f93c843ac40686a7c3fa514060a97d9de93e5e72b"}, - {file = "numpy-2.3.0-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:d344ca32ab482bcf8735d8f95091ad081f97120546f3d250240868430ce52555"}, - {file = "numpy-2.3.0-cp313-cp313t-win32.whl", hash = "sha256:48a2e8eaf76364c32a1feaa60d6925eaf32ed7a040183b807e02674305beef61"}, - {file = "numpy-2.3.0-cp313-cp313t-win_amd64.whl", hash = "sha256:ba17f93a94e503551f154de210e4d50c5e3ee20f7e7a1b5f6ce3f22d419b93bb"}, - {file = "numpy-2.3.0-cp313-cp313t-win_arm64.whl", hash = "sha256:f14e016d9409680959691c109be98c436c6249eaf7f118b424679793607b5944"}, - {file = "numpy-2.3.0-pp311-pypy311_pp73-macosx_10_15_x86_64.whl", hash = "sha256:80b46117c7359de8167cc00a2c7d823bdd505e8c7727ae0871025a86d668283b"}, - {file = "numpy-2.3.0-pp311-pypy311_pp73-macosx_14_0_arm64.whl", hash = "sha256:5814a0f43e70c061f47abd5857d120179609ddc32a613138cbb6c4e9e2dbdda5"}, - {file = "numpy-2.3.0-pp311-pypy311_pp73-macosx_14_0_x86_64.whl", hash = "sha256:ef6c1e88fd6b81ac6d215ed71dc8cd027e54d4bf1d2682d362449097156267a2"}, - {file = "numpy-2.3.0-pp311-pypy311_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:33a5a12a45bb82d9997e2c0b12adae97507ad7c347546190a18ff14c28bbca12"}, - {file = "numpy-2.3.0-pp311-pypy311_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:54dfc8681c1906d239e95ab1508d0a533c4a9505e52ee2d71a5472b04437ef97"}, - {file = "numpy-2.3.0-pp311-pypy311_pp73-win_amd64.whl", hash = "sha256:e017a8a251ff4d18d71f139e28bdc7c31edba7a507f72b1414ed902cbe48c74d"}, - {file = "numpy-2.3.0.tar.gz", hash = "sha256:581f87f9e9e9db2cba2141400e160e9dd644ee248788d6f90636eeb8fd9260a6"}, + {file = "numpy-2.3.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:6ea9e48336a402551f52cd8f593343699003d2353daa4b72ce8d34f66b722070"}, + {file = "numpy-2.3.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5ccb7336eaf0e77c1635b232c141846493a588ec9ea777a7c24d7166bb8533ae"}, + {file = "numpy-2.3.1-cp311-cp311-macosx_14_0_arm64.whl", hash = "sha256:0bb3a4a61e1d327e035275d2a993c96fa786e4913aa089843e6a2d9dd205c66a"}, + {file = "numpy-2.3.1-cp311-cp311-macosx_14_0_x86_64.whl", hash = "sha256:e344eb79dab01f1e838ebb67aab09965fb271d6da6b00adda26328ac27d4a66e"}, + {file = "numpy-2.3.1-cp311-cp311-manylinux_2_28_aarch64.whl", hash = "sha256:467db865b392168ceb1ef1ffa6f5a86e62468c43e0cfb4ab6da667ede10e58db"}, + {file = "numpy-2.3.1-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:afed2ce4a84f6b0fc6c1ce734ff368cbf5a5e24e8954a338f3bdffa0718adffb"}, + {file = "numpy-2.3.1-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:0025048b3c1557a20bc80d06fdeb8cc7fc193721484cca82b2cfa072fec71a93"}, + {file = "numpy-2.3.1-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:a5ee121b60aa509679b682819c602579e1df14a5b07fe95671c8849aad8f2115"}, + {file = "numpy-2.3.1-cp311-cp311-win32.whl", hash = "sha256:a8b740f5579ae4585831b3cf0e3b0425c667274f82a484866d2adf9570539369"}, + {file = "numpy-2.3.1-cp311-cp311-win_amd64.whl", hash = "sha256:d4580adadc53311b163444f877e0789f1c8861e2698f6b2a4ca852fda154f3ff"}, + {file = "numpy-2.3.1-cp311-cp311-win_arm64.whl", hash = "sha256:ec0bdafa906f95adc9a0c6f26a4871fa753f25caaa0e032578a30457bff0af6a"}, + {file = "numpy-2.3.1-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:2959d8f268f3d8ee402b04a9ec4bb7604555aeacf78b360dc4ec27f1d508177d"}, + {file = "numpy-2.3.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:762e0c0c6b56bdedfef9a8e1d4538556438288c4276901ea008ae44091954e29"}, + {file = "numpy-2.3.1-cp312-cp312-macosx_14_0_arm64.whl", hash = "sha256:867ef172a0976aaa1f1d1b63cf2090de8b636a7674607d514505fb7276ab08fc"}, + {file = "numpy-2.3.1-cp312-cp312-macosx_14_0_x86_64.whl", hash = "sha256:4e602e1b8682c2b833af89ba641ad4176053aaa50f5cacda1a27004352dde943"}, + {file = "numpy-2.3.1-cp312-cp312-manylinux_2_28_aarch64.whl", hash = "sha256:8e333040d069eba1652fb08962ec5b76af7f2c7bce1df7e1418c8055cf776f25"}, + {file = "numpy-2.3.1-cp312-cp312-manylinux_2_28_x86_64.whl", hash = "sha256:e7cbf5a5eafd8d230a3ce356d892512185230e4781a361229bd902ff403bc660"}, + {file = "numpy-2.3.1-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:5f1b8f26d1086835f442286c1d9b64bb3974b0b1e41bb105358fd07d20872952"}, + {file = "numpy-2.3.1-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:ee8340cb48c9b7a5899d1149eece41ca535513a9698098edbade2a8e7a84da77"}, + {file = "numpy-2.3.1-cp312-cp312-win32.whl", hash = "sha256:e772dda20a6002ef7061713dc1e2585bc1b534e7909b2030b5a46dae8ff077ab"}, + {file = "numpy-2.3.1-cp312-cp312-win_amd64.whl", hash = "sha256:cfecc7822543abdea6de08758091da655ea2210b8ffa1faf116b940693d3df76"}, + {file = "numpy-2.3.1-cp312-cp312-win_arm64.whl", hash = "sha256:7be91b2239af2658653c5bb6f1b8bccafaf08226a258caf78ce44710a0160d30"}, + {file = "numpy-2.3.1-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:25a1992b0a3fdcdaec9f552ef10d8103186f5397ab45e2d25f8ac51b1a6b97e8"}, + {file = "numpy-2.3.1-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:7dea630156d39b02a63c18f508f85010230409db5b2927ba59c8ba4ab3e8272e"}, + {file = "numpy-2.3.1-cp313-cp313-macosx_14_0_arm64.whl", hash = "sha256:bada6058dd886061f10ea15f230ccf7dfff40572e99fef440a4a857c8728c9c0"}, + {file = "numpy-2.3.1-cp313-cp313-macosx_14_0_x86_64.whl", hash = "sha256:a894f3816eb17b29e4783e5873f92faf55b710c2519e5c351767c51f79d8526d"}, + {file = "numpy-2.3.1-cp313-cp313-manylinux_2_28_aarch64.whl", hash = "sha256:18703df6c4a4fee55fd3d6e5a253d01c5d33a295409b03fda0c86b3ca2ff41a1"}, + {file = "numpy-2.3.1-cp313-cp313-manylinux_2_28_x86_64.whl", hash = "sha256:5902660491bd7a48b2ec16c23ccb9124b8abfd9583c5fdfa123fe6b421e03de1"}, + {file = "numpy-2.3.1-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:36890eb9e9d2081137bd78d29050ba63b8dab95dff7912eadf1185e80074b2a0"}, + {file = "numpy-2.3.1-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:a780033466159c2270531e2b8ac063704592a0bc62ec4a1b991c7c40705eb0e8"}, + {file = "numpy-2.3.1-cp313-cp313-win32.whl", hash = "sha256:39bff12c076812595c3a306f22bfe49919c5513aa1e0e70fac756a0be7c2a2b8"}, + {file = "numpy-2.3.1-cp313-cp313-win_amd64.whl", hash = "sha256:8d5ee6eec45f08ce507a6570e06f2f879b374a552087a4179ea7838edbcbfa42"}, + {file = "numpy-2.3.1-cp313-cp313-win_arm64.whl", hash = "sha256:0c4d9e0a8368db90f93bd192bfa771ace63137c3488d198ee21dfb8e7771916e"}, + {file = "numpy-2.3.1-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:b0b5397374f32ec0649dd98c652a1798192042e715df918c20672c62fb52d4b8"}, + {file = "numpy-2.3.1-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:c5bdf2015ccfcee8253fb8be695516ac4457c743473a43290fd36eba6a1777eb"}, + {file = "numpy-2.3.1-cp313-cp313t-macosx_14_0_arm64.whl", hash = "sha256:d70f20df7f08b90a2062c1f07737dd340adccf2068d0f1b9b3d56e2038979fee"}, + {file = "numpy-2.3.1-cp313-cp313t-macosx_14_0_x86_64.whl", hash = "sha256:2fb86b7e58f9ac50e1e9dd1290154107e47d1eef23a0ae9145ded06ea606f992"}, + {file = "numpy-2.3.1-cp313-cp313t-manylinux_2_28_aarch64.whl", hash = "sha256:23ab05b2d241f76cb883ce8b9a93a680752fbfcbd51c50eff0b88b979e471d8c"}, + {file = "numpy-2.3.1-cp313-cp313t-manylinux_2_28_x86_64.whl", hash = "sha256:ce2ce9e5de4703a673e705183f64fd5da5bf36e7beddcb63a25ee2286e71ca48"}, + {file = "numpy-2.3.1-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:c4913079974eeb5c16ccfd2b1f09354b8fed7e0d6f2cab933104a09a6419b1ee"}, + {file = "numpy-2.3.1-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:010ce9b4f00d5c036053ca684c77441f2f2c934fd23bee058b4d6f196efd8280"}, + {file = "numpy-2.3.1-cp313-cp313t-win32.whl", hash = "sha256:6269b9edfe32912584ec496d91b00b6d34282ca1d07eb10e82dfc780907d6c2e"}, + {file = "numpy-2.3.1-cp313-cp313t-win_amd64.whl", hash = "sha256:2a809637460e88a113e186e87f228d74ae2852a2e0c44de275263376f17b5bdc"}, + {file = "numpy-2.3.1-cp313-cp313t-win_arm64.whl", hash = "sha256:eccb9a159db9aed60800187bc47a6d3451553f0e1b08b068d8b277ddfbb9b244"}, + {file = "numpy-2.3.1-pp311-pypy311_pp73-macosx_10_15_x86_64.whl", hash = "sha256:ad506d4b09e684394c42c966ec1527f6ebc25da7f4da4b1b056606ffe446b8a3"}, + {file = "numpy-2.3.1-pp311-pypy311_pp73-macosx_14_0_arm64.whl", hash = "sha256:ebb8603d45bc86bbd5edb0d63e52c5fd9e7945d3a503b77e486bd88dde67a19b"}, + {file = "numpy-2.3.1-pp311-pypy311_pp73-macosx_14_0_x86_64.whl", hash = "sha256:15aa4c392ac396e2ad3d0a2680c0f0dee420f9fed14eef09bdb9450ee6dcb7b7"}, + {file = "numpy-2.3.1-pp311-pypy311_pp73-manylinux_2_28_aarch64.whl", hash = "sha256:c6e0bf9d1a2f50d2b65a7cf56db37c095af17b59f6c132396f7c6d5dd76484df"}, + {file = "numpy-2.3.1-pp311-pypy311_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:eabd7e8740d494ce2b4ea0ff05afa1b7b291e978c0ae075487c51e8bd93c0c68"}, + {file = "numpy-2.3.1-pp311-pypy311_pp73-win_amd64.whl", hash = "sha256:e610832418a2bc09d974cc9fecebfa51e9532d6190223bc5ef6a7402ebf3b5cb"}, + {file = "numpy-2.3.1.tar.gz", hash = "sha256:1ec9ae20a4226da374362cca3c62cd753faf2f951440b0e3b98e93c235441d2b"}, ] [[package]] name = "oauthlib" -version = "3.2.2" +version = "3.3.1" description = "A generic, spec-compliant, thorough implementation of the OAuth request-signing logic" optional = false -python-versions = ">=3.6" +python-versions = ">=3.8" groups = ["main"] files = [ - {file = "oauthlib-3.2.2-py3-none-any.whl", hash = "sha256:8139f29aac13e25d502680e9e19963e83f16838d48a0d71c287fe40e7067fbca"}, - {file = "oauthlib-3.2.2.tar.gz", hash = "sha256:9859c40929662bec5d64f34d01c99e093149682a3f38915dc0655d5a633dd918"}, + {file = "oauthlib-3.3.1-py3-none-any.whl", hash = "sha256:88119c938d2b8fb88561af5f6ee0eec8cc8d552b7bb1f712743136eb7523b7a1"}, + {file = "oauthlib-3.3.1.tar.gz", hash = "sha256:0f0f8aa759826a193cf66c12ea1af1637f87b9b4622d46e866952bb022e538c9"}, ] [package.extras] @@ -2562,54 +2597,54 @@ files = [ [[package]] name = "pandas" -version = "2.3.0" +version = "2.3.1" description = "Powerful data structures for data analysis, time series, and statistics" optional = false python-versions = ">=3.9" groups = ["main"] files = [ - {file = "pandas-2.3.0-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:625466edd01d43b75b1883a64d859168e4556261a5035b32f9d743b67ef44634"}, - {file = "pandas-2.3.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:a6872d695c896f00df46b71648eea332279ef4077a409e2fe94220208b6bb675"}, - {file = "pandas-2.3.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:f4dd97c19bd06bc557ad787a15b6489d2614ddaab5d104a0310eb314c724b2d2"}, - {file = "pandas-2.3.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:034abd6f3db8b9880aaee98f4f5d4dbec7c4829938463ec046517220b2f8574e"}, - {file = "pandas-2.3.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:23c2b2dc5213810208ca0b80b8666670eb4660bbfd9d45f58592cc4ddcfd62e1"}, - {file = "pandas-2.3.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:39ff73ec07be5e90330cc6ff5705c651ace83374189dcdcb46e6ff54b4a72cd6"}, - {file = "pandas-2.3.0-cp310-cp310-win_amd64.whl", hash = "sha256:40cecc4ea5abd2921682b57532baea5588cc5f80f0231c624056b146887274d2"}, - {file = "pandas-2.3.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:8adff9f138fc614347ff33812046787f7d43b3cef7c0f0171b3340cae333f6ca"}, - {file = "pandas-2.3.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:e5f08eb9a445d07720776df6e641975665c9ea12c9d8a331e0f6890f2dcd76ef"}, - {file = "pandas-2.3.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:fa35c266c8cd1a67d75971a1912b185b492d257092bdd2709bbdebe574ed228d"}, - {file = "pandas-2.3.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:14a0cc77b0f089d2d2ffe3007db58f170dae9b9f54e569b299db871a3ab5bf46"}, - {file = "pandas-2.3.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:c06f6f144ad0a1bf84699aeea7eff6068ca5c63ceb404798198af7eb86082e33"}, - {file = "pandas-2.3.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:ed16339bc354a73e0a609df36d256672c7d296f3f767ac07257801aa064ff73c"}, - {file = "pandas-2.3.0-cp311-cp311-win_amd64.whl", hash = "sha256:fa07e138b3f6c04addfeaf56cc7fdb96c3b68a3fe5e5401251f231fce40a0d7a"}, - {file = "pandas-2.3.0-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:2eb4728a18dcd2908c7fccf74a982e241b467d178724545a48d0caf534b38ebf"}, - {file = "pandas-2.3.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:b9d8c3187be7479ea5c3d30c32a5d73d62a621166675063b2edd21bc47614027"}, - {file = "pandas-2.3.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9ff730713d4c4f2f1c860e36c005c7cefc1c7c80c21c0688fd605aa43c9fcf09"}, - {file = "pandas-2.3.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ba24af48643b12ffe49b27065d3babd52702d95ab70f50e1b34f71ca703e2c0d"}, - {file = "pandas-2.3.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:404d681c698e3c8a40a61d0cd9412cc7364ab9a9cc6e144ae2992e11a2e77a20"}, - {file = "pandas-2.3.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:6021910b086b3ca756755e86ddc64e0ddafd5e58e076c72cb1585162e5ad259b"}, - {file = "pandas-2.3.0-cp312-cp312-win_amd64.whl", hash = "sha256:094e271a15b579650ebf4c5155c05dcd2a14fd4fdd72cf4854b2f7ad31ea30be"}, - {file = "pandas-2.3.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:2c7e2fc25f89a49a11599ec1e76821322439d90820108309bf42130d2f36c983"}, - {file = "pandas-2.3.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:c6da97aeb6a6d233fb6b17986234cc723b396b50a3c6804776351994f2a658fd"}, - {file = "pandas-2.3.0-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bb32dc743b52467d488e7a7c8039b821da2826a9ba4f85b89ea95274f863280f"}, - {file = "pandas-2.3.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:213cd63c43263dbb522c1f8a7c9d072e25900f6975596f883f4bebd77295d4f3"}, - {file = "pandas-2.3.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:1d2b33e68d0ce64e26a4acc2e72d747292084f4e8db4c847c6f5f6cbe56ed6d8"}, - {file = "pandas-2.3.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:430a63bae10b5086995db1b02694996336e5a8ac9a96b4200572b413dfdfccb9"}, - {file = "pandas-2.3.0-cp313-cp313-win_amd64.whl", hash = "sha256:4930255e28ff5545e2ca404637bcc56f031893142773b3468dc021c6c32a1390"}, - {file = "pandas-2.3.0-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:f925f1ef673b4bd0271b1809b72b3270384f2b7d9d14a189b12b7fc02574d575"}, - {file = "pandas-2.3.0-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:e78ad363ddb873a631e92a3c063ade1ecfb34cae71e9a2be6ad100f875ac1042"}, - {file = "pandas-2.3.0-cp313-cp313t-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:951805d146922aed8357e4cc5671b8b0b9be1027f0619cea132a9f3f65f2f09c"}, - {file = "pandas-2.3.0-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1a881bc1309f3fce34696d07b00f13335c41f5f5a8770a33b09ebe23261cfc67"}, - {file = "pandas-2.3.0-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:e1991bbb96f4050b09b5f811253c4f3cf05ee89a589379aa36cd623f21a31d6f"}, - {file = "pandas-2.3.0-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:bb3be958022198531eb7ec2008cfc78c5b1eed51af8600c6c5d9160d89d8d249"}, - {file = "pandas-2.3.0-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:9efc0acbbffb5236fbdf0409c04edce96bec4bdaa649d49985427bd1ec73e085"}, - {file = "pandas-2.3.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:75651c14fde635e680496148a8526b328e09fe0572d9ae9b638648c46a544ba3"}, - {file = "pandas-2.3.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:bf5be867a0541a9fb47a4be0c5790a4bccd5b77b92f0a59eeec9375fafc2aa14"}, - {file = "pandas-2.3.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:84141f722d45d0c2a89544dd29d35b3abfc13d2250ed7e68394eda7564bd6324"}, - {file = "pandas-2.3.0-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:f95a2aef32614ed86216d3c450ab12a4e82084e8102e355707a1d96e33d51c34"}, - {file = "pandas-2.3.0-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:e0f51973ba93a9f97185049326d75b942b9aeb472bec616a129806facb129ebb"}, - {file = "pandas-2.3.0-cp39-cp39-win_amd64.whl", hash = "sha256:b198687ca9c8529662213538a9bb1e60fa0bf0f6af89292eb68fea28743fcd5a"}, - {file = "pandas-2.3.0.tar.gz", hash = "sha256:34600ab34ebf1131a7613a260a61dbe8b62c188ec0ea4c296da7c9a06b004133"}, + {file = "pandas-2.3.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:22c2e866f7209ebc3a8f08d75766566aae02bcc91d196935a1d9e59c7b990ac9"}, + {file = "pandas-2.3.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:3583d348546201aff730c8c47e49bc159833f971c2899d6097bce68b9112a4f1"}, + {file = "pandas-2.3.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:0f951fbb702dacd390561e0ea45cdd8ecfa7fb56935eb3dd78e306c19104b9b0"}, + {file = "pandas-2.3.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:cd05b72ec02ebfb993569b4931b2e16fbb4d6ad6ce80224a3ee838387d83a191"}, + {file = "pandas-2.3.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:1b916a627919a247d865aed068eb65eb91a344b13f5b57ab9f610b7716c92de1"}, + {file = "pandas-2.3.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:fe67dc676818c186d5a3d5425250e40f179c2a89145df477dd82945eaea89e97"}, + {file = "pandas-2.3.1-cp310-cp310-win_amd64.whl", hash = "sha256:2eb789ae0274672acbd3c575b0598d213345660120a257b47b5dafdc618aec83"}, + {file = "pandas-2.3.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:2b0540963d83431f5ce8870ea02a7430adca100cec8a050f0811f8e31035541b"}, + {file = "pandas-2.3.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:fe7317f578c6a153912bd2292f02e40c1d8f253e93c599e82620c7f69755c74f"}, + {file = "pandas-2.3.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e6723a27ad7b244c0c79d8e7007092d7c8f0f11305770e2f4cd778b3ad5f9f85"}, + {file = "pandas-2.3.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:3462c3735fe19f2638f2c3a40bd94ec2dc5ba13abbb032dd2fa1f540a075509d"}, + {file = "pandas-2.3.1-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:98bcc8b5bf7afed22cc753a28bc4d9e26e078e777066bc53fac7904ddef9a678"}, + {file = "pandas-2.3.1-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:4d544806b485ddf29e52d75b1f559142514e60ef58a832f74fb38e48d757b299"}, + {file = "pandas-2.3.1-cp311-cp311-win_amd64.whl", hash = "sha256:b3cd4273d3cb3707b6fffd217204c52ed92859533e31dc03b7c5008aa933aaab"}, + {file = "pandas-2.3.1-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:689968e841136f9e542020698ee1c4fbe9caa2ed2213ae2388dc7b81721510d3"}, + {file = "pandas-2.3.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:025e92411c16cbe5bb2a4abc99732a6b132f439b8aab23a59fa593eb00704232"}, + {file = "pandas-2.3.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:9b7ff55f31c4fcb3e316e8f7fa194566b286d6ac430afec0d461163312c5841e"}, + {file = "pandas-2.3.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7dcb79bf373a47d2a40cf7232928eb7540155abbc460925c2c96d2d30b006eb4"}, + {file = "pandas-2.3.1-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:56a342b231e8862c96bdb6ab97170e203ce511f4d0429589c8ede1ee8ece48b8"}, + {file = "pandas-2.3.1-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:ca7ed14832bce68baef331f4d7f294411bed8efd032f8109d690df45e00c4679"}, + {file = "pandas-2.3.1-cp312-cp312-win_amd64.whl", hash = "sha256:ac942bfd0aca577bef61f2bc8da8147c4ef6879965ef883d8e8d5d2dc3e744b8"}, + {file = "pandas-2.3.1-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:9026bd4a80108fac2239294a15ef9003c4ee191a0f64b90f170b40cfb7cf2d22"}, + {file = "pandas-2.3.1-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:6de8547d4fdb12421e2d047a2c446c623ff4c11f47fddb6b9169eb98ffba485a"}, + {file = "pandas-2.3.1-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:782647ddc63c83133b2506912cc6b108140a38a37292102aaa19c81c83db2928"}, + {file = "pandas-2.3.1-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2ba6aff74075311fc88504b1db890187a3cd0f887a5b10f5525f8e2ef55bfdb9"}, + {file = "pandas-2.3.1-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:e5635178b387bd2ba4ac040f82bc2ef6e6b500483975c4ebacd34bec945fda12"}, + {file = "pandas-2.3.1-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:6f3bf5ec947526106399a9e1d26d40ee2b259c66422efdf4de63c848492d91bb"}, + {file = "pandas-2.3.1-cp313-cp313-win_amd64.whl", hash = "sha256:1c78cf43c8fde236342a1cb2c34bcff89564a7bfed7e474ed2fffa6aed03a956"}, + {file = "pandas-2.3.1-cp313-cp313t-macosx_10_13_x86_64.whl", hash = "sha256:8dfc17328e8da77be3cf9f47509e5637ba8f137148ed0e9b5241e1baf526e20a"}, + {file = "pandas-2.3.1-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:ec6c851509364c59a5344458ab935e6451b31b818be467eb24b0fe89bd05b6b9"}, + {file = "pandas-2.3.1-cp313-cp313t-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:911580460fc4884d9b05254b38a6bfadddfcc6aaef856fb5859e7ca202e45275"}, + {file = "pandas-2.3.1-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2f4d6feeba91744872a600e6edbbd5b033005b431d5ae8379abee5bcfa479fab"}, + {file = "pandas-2.3.1-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:fe37e757f462d31a9cd7580236a82f353f5713a80e059a29753cf938c6775d96"}, + {file = "pandas-2.3.1-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:5db9637dbc24b631ff3707269ae4559bce4b7fd75c1c4d7e13f40edc42df4444"}, + {file = "pandas-2.3.1-cp39-cp39-macosx_10_9_x86_64.whl", hash = "sha256:4645f770f98d656f11c69e81aeb21c6fca076a44bed3dcbb9396a4311bc7f6d8"}, + {file = "pandas-2.3.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:342e59589cc454aaff7484d75b816a433350b3d7964d7847327edda4d532a2e3"}, + {file = "pandas-2.3.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1d12f618d80379fde6af007f65f0c25bd3e40251dbd1636480dfffce2cf1e6da"}, + {file = "pandas-2.3.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:dd71c47a911da120d72ef173aeac0bf5241423f9bfea57320110a978457e069e"}, + {file = "pandas-2.3.1-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:09e3b1587f0f3b0913e21e8b32c3119174551deb4a4eba4a89bc7377947977e7"}, + {file = "pandas-2.3.1-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:2323294c73ed50f612f67e2bf3ae45aea04dce5690778e08a09391897f35ff88"}, + {file = "pandas-2.3.1-cp39-cp39-win_amd64.whl", hash = "sha256:b4b0de34dc8499c2db34000ef8baad684cfa4cbd836ecee05f323ebfba348c7d"}, + {file = "pandas-2.3.1.tar.gz", hash = "sha256:0a95b9ac964fe83ce317827f80304d37388ea77616b1425f0ae41c9d2d0d7bb2"}, ] [package.dependencies] @@ -2735,7 +2770,7 @@ version = "1.6.0" description = "plugin and hook calling mechanisms for python" optional = false python-versions = ">=3.9" -groups = ["test"] +groups = ["dev", "test"] files = [ {file = "pluggy-1.6.0-py3-none-any.whl", hash = "sha256:e920276dd6813095e9377c0bc5566d94c932c33b27a3e3945d8389c374dd4746"}, {file = "pluggy-1.6.0.tar.gz", hash = "sha256:7dcc130b76258d33b90f61b658791dede3486c3e6bfb003ee5c9bfb396dd22f3"}, @@ -3287,14 +3322,14 @@ typing-extensions = ">=4.6.0,<4.7.0 || >4.7.0" [[package]] name = "pygments" -version = "2.19.1" +version = "2.19.2" description = "Pygments is a syntax highlighting package written in Python." optional = false python-versions = ">=3.8" -groups = ["main", "docs", "test"] +groups = ["main", "dev", "docs", "test"] files = [ - {file = "pygments-2.19.1-py3-none-any.whl", hash = "sha256:9ea1544ad55cecf4b8242fab6dd35a93bbce657034b0611ee383099054ab6d8c"}, - {file = "pygments-2.19.1.tar.gz", hash = "sha256:61c16d2a8576dc0649d9f39e089b5f02bcd27fba10d8fb4dcc28173f7a45151f"}, + {file = "pygments-2.19.2-py3-none-any.whl", hash = "sha256:86540386c03d588bb81d44bc3928634ff26449851e99741617ecb9037ee5ec0b"}, + {file = "pygments-2.19.2.tar.gz", hash = "sha256:636cb2477cec7f8952536970bc533bc43743542f70392ae026374600add5b887"}, ] [package.extras] @@ -3381,21 +3416,46 @@ files = [ [package.dependencies] pytest = ">=7.0.0" +[[package]] +name = "python-client" +version = "0.0.0-dev" +description = "Python Client for Kuberay" +optional = false +python-versions = "^3.11" +groups = ["main"] +files = [] +develop = false + +[package.dependencies] +kubernetes = ">=25.0.0" + +[package.source] +type = "git" +url = "https://github.com/ray-project/kuberay.git" +reference = "b2fd91b58c2bbe22f9b4f730c5a8f3180c05e570" +resolved_reference = "b2fd91b58c2bbe22f9b4f730c5a8f3180c05e570" +subdirectory = "clients/python-client" + [[package]] name = "python-dateutil" -version = "2.9.0.post0" +version = "3.9.0" description = "Extensions to the standard Python datetime module" optional = false python-versions = "!=3.0.*,!=3.1.*,!=3.2.*,>=2.7" groups = ["main", "test"] files = [ - {file = "python-dateutil-2.9.0.post0.tar.gz", hash = "sha256:37dd54208da7e1cd875388217d5e00ebd4179249f90fb72437e91a35459a0ad3"}, - {file = "python_dateutil-2.9.0.post0-py2.py3-none-any.whl", hash = "sha256:a8b2bc7bffae282281c8140a97d3aa9c14da0b136dfe83f850eea9a5f7470427"}, + {file = "python-dateutil-3.9.0.tar.gz", hash = "sha256:e090c9a06b858a55d8b6a518fc54d079646eb7262b373ff98f8f13877a5327ec"}, + {file = "python_dateutil-3.9.0-py2.py3-none-any.whl", hash = "sha256:971787138d3cb47d927800e544872edc9e49f33ad1335adc139c409aa5e6a9a8"}, ] [package.dependencies] six = ">=1.5" +[package.source] +type = "legacy" +url = "https://test.pypi.org/simple" +reference = "testpypi" + [[package]] name = "python-json-logger" version = "3.3.0" @@ -3533,105 +3593,91 @@ files = [ [[package]] name = "pyzmq" -version = "26.4.0" +version = "27.0.0" description = "Python bindings for 0MQ" optional = false python-versions = ">=3.8" groups = ["test"] files = [ - {file = "pyzmq-26.4.0-cp310-cp310-macosx_10_15_universal2.whl", hash = "sha256:0329bdf83e170ac133f44a233fc651f6ed66ef8e66693b5af7d54f45d1ef5918"}, - {file = "pyzmq-26.4.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:398a825d2dea96227cf6460ce0a174cf7657d6f6827807d4d1ae9d0f9ae64315"}, - {file = "pyzmq-26.4.0-cp310-cp310-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:6d52d62edc96787f5c1dfa6c6ccff9b581cfae5a70d94ec4c8da157656c73b5b"}, - {file = "pyzmq-26.4.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1410c3a3705db68d11eb2424d75894d41cff2f64d948ffe245dd97a9debfebf4"}, - {file = "pyzmq-26.4.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:7dacb06a9c83b007cc01e8e5277f94c95c453c5851aac5e83efe93e72226353f"}, - {file = "pyzmq-26.4.0-cp310-cp310-musllinux_1_1_aarch64.whl", hash = "sha256:6bab961c8c9b3a4dc94d26e9b2cdf84de9918931d01d6ff38c721a83ab3c0ef5"}, - {file = "pyzmq-26.4.0-cp310-cp310-musllinux_1_1_i686.whl", hash = "sha256:7a5c09413b924d96af2aa8b57e76b9b0058284d60e2fc3730ce0f979031d162a"}, - {file = "pyzmq-26.4.0-cp310-cp310-musllinux_1_1_x86_64.whl", hash = "sha256:7d489ac234d38e57f458fdbd12a996bfe990ac028feaf6f3c1e81ff766513d3b"}, - {file = "pyzmq-26.4.0-cp310-cp310-win32.whl", hash = "sha256:dea1c8db78fb1b4b7dc9f8e213d0af3fc8ecd2c51a1d5a3ca1cde1bda034a980"}, - {file = "pyzmq-26.4.0-cp310-cp310-win_amd64.whl", hash = "sha256:fa59e1f5a224b5e04dc6c101d7186058efa68288c2d714aa12d27603ae93318b"}, - {file = "pyzmq-26.4.0-cp310-cp310-win_arm64.whl", hash = "sha256:a651fe2f447672f4a815e22e74630b6b1ec3a1ab670c95e5e5e28dcd4e69bbb5"}, - {file = "pyzmq-26.4.0-cp311-cp311-macosx_10_15_universal2.whl", hash = "sha256:bfcf82644c9b45ddd7cd2a041f3ff8dce4a0904429b74d73a439e8cab1bd9e54"}, - {file = "pyzmq-26.4.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e9bcae3979b2654d5289d3490742378b2f3ce804b0b5fd42036074e2bf35b030"}, - {file = "pyzmq-26.4.0-cp311-cp311-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ccdff8ac4246b6fb60dcf3982dfaeeff5dd04f36051fe0632748fc0aa0679c01"}, - {file = "pyzmq-26.4.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4550af385b442dc2d55ab7717837812799d3674cb12f9a3aa897611839c18e9e"}, - {file = "pyzmq-26.4.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:2f9f7ffe9db1187a253fca95191854b3fda24696f086e8789d1d449308a34b88"}, - {file = "pyzmq-26.4.0-cp311-cp311-musllinux_1_1_aarch64.whl", hash = "sha256:3709c9ff7ba61589b7372923fd82b99a81932b592a5c7f1a24147c91da9a68d6"}, - {file = "pyzmq-26.4.0-cp311-cp311-musllinux_1_1_i686.whl", hash = "sha256:f8f3c30fb2d26ae5ce36b59768ba60fb72507ea9efc72f8f69fa088450cff1df"}, - {file = "pyzmq-26.4.0-cp311-cp311-musllinux_1_1_x86_64.whl", hash = "sha256:382a4a48c8080e273427fc692037e3f7d2851959ffe40864f2db32646eeb3cef"}, - {file = "pyzmq-26.4.0-cp311-cp311-win32.whl", hash = "sha256:d56aad0517d4c09e3b4f15adebba8f6372c5102c27742a5bdbfc74a7dceb8fca"}, - {file = "pyzmq-26.4.0-cp311-cp311-win_amd64.whl", hash = "sha256:963977ac8baed7058c1e126014f3fe58b3773f45c78cce7af5c26c09b6823896"}, - {file = "pyzmq-26.4.0-cp311-cp311-win_arm64.whl", hash = "sha256:c0c8e8cadc81e44cc5088fcd53b9b3b4ce9344815f6c4a03aec653509296fae3"}, - {file = "pyzmq-26.4.0-cp312-cp312-macosx_10_15_universal2.whl", hash = "sha256:5227cb8da4b6f68acfd48d20c588197fd67745c278827d5238c707daf579227b"}, - {file = "pyzmq-26.4.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e1c07a7fa7f7ba86554a2b1bef198c9fed570c08ee062fd2fd6a4dcacd45f905"}, - {file = "pyzmq-26.4.0-cp312-cp312-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:ae775fa83f52f52de73183f7ef5395186f7105d5ed65b1ae65ba27cb1260de2b"}, - {file = "pyzmq-26.4.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:66c760d0226ebd52f1e6b644a9e839b5db1e107a23f2fcd46ec0569a4fdd4e63"}, - {file = "pyzmq-26.4.0-cp312-cp312-manylinux_2_28_x86_64.whl", hash = "sha256:ef8c6ecc1d520debc147173eaa3765d53f06cd8dbe7bd377064cdbc53ab456f5"}, - {file = "pyzmq-26.4.0-cp312-cp312-musllinux_1_1_aarch64.whl", hash = "sha256:3150ef4084e163dec29ae667b10d96aad309b668fac6810c9e8c27cf543d6e0b"}, - {file = "pyzmq-26.4.0-cp312-cp312-musllinux_1_1_i686.whl", hash = "sha256:4448c9e55bf8329fa1dcedd32f661bf611214fa70c8e02fee4347bc589d39a84"}, - {file = "pyzmq-26.4.0-cp312-cp312-musllinux_1_1_x86_64.whl", hash = "sha256:e07dde3647afb084d985310d067a3efa6efad0621ee10826f2cb2f9a31b89d2f"}, - {file = "pyzmq-26.4.0-cp312-cp312-win32.whl", hash = "sha256:ba034a32ecf9af72adfa5ee383ad0fd4f4e38cdb62b13624278ef768fe5b5b44"}, - {file = "pyzmq-26.4.0-cp312-cp312-win_amd64.whl", hash = "sha256:056a97aab4064f526ecb32f4343917a4022a5d9efb6b9df990ff72e1879e40be"}, - {file = "pyzmq-26.4.0-cp312-cp312-win_arm64.whl", hash = "sha256:2f23c750e485ce1eb639dbd576d27d168595908aa2d60b149e2d9e34c9df40e0"}, - {file = "pyzmq-26.4.0-cp313-cp313-macosx_10_15_universal2.whl", hash = "sha256:c43fac689880f5174d6fc864857d1247fe5cfa22b09ed058a344ca92bf5301e3"}, - {file = "pyzmq-26.4.0-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:902aca7eba477657c5fb81c808318460328758e8367ecdd1964b6330c73cae43"}, - {file = "pyzmq-26.4.0-cp313-cp313-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:e5e48a830bfd152fe17fbdeaf99ac5271aa4122521bf0d275b6b24e52ef35eb6"}, - {file = "pyzmq-26.4.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:31be2b6de98c824c06f5574331f805707c667dc8f60cb18580b7de078479891e"}, - {file = "pyzmq-26.4.0-cp313-cp313-manylinux_2_28_x86_64.whl", hash = "sha256:6332452034be001bbf3206ac59c0d2a7713de5f25bb38b06519fc6967b7cf771"}, - {file = "pyzmq-26.4.0-cp313-cp313-musllinux_1_1_aarch64.whl", hash = "sha256:da8c0f5dd352136853e6a09b1b986ee5278dfddfebd30515e16eae425c872b30"}, - {file = "pyzmq-26.4.0-cp313-cp313-musllinux_1_1_i686.whl", hash = "sha256:f4ccc1a0a2c9806dda2a2dd118a3b7b681e448f3bb354056cad44a65169f6d86"}, - {file = "pyzmq-26.4.0-cp313-cp313-musllinux_1_1_x86_64.whl", hash = "sha256:1c0b5fceadbab461578daf8d1dcc918ebe7ddd2952f748cf30c7cf2de5d51101"}, - {file = "pyzmq-26.4.0-cp313-cp313-win32.whl", hash = "sha256:28e2b0ff5ba4b3dd11062d905682bad33385cfa3cc03e81abd7f0822263e6637"}, - {file = "pyzmq-26.4.0-cp313-cp313-win_amd64.whl", hash = "sha256:23ecc9d241004c10e8b4f49d12ac064cd7000e1643343944a10df98e57bc544b"}, - {file = "pyzmq-26.4.0-cp313-cp313-win_arm64.whl", hash = "sha256:1edb0385c7f025045d6e0f759d4d3afe43c17a3d898914ec6582e6f464203c08"}, - {file = "pyzmq-26.4.0-cp313-cp313t-macosx_10_15_universal2.whl", hash = "sha256:93a29e882b2ba1db86ba5dd5e88e18e0ac6b627026c5cfbec9983422011b82d4"}, - {file = "pyzmq-26.4.0-cp313-cp313t-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:cb45684f276f57110bb89e4300c00f1233ca631f08f5f42528a5c408a79efc4a"}, - {file = "pyzmq-26.4.0-cp313-cp313t-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:f72073e75260cb301aad4258ad6150fa7f57c719b3f498cb91e31df16784d89b"}, - {file = "pyzmq-26.4.0-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:be37e24b13026cfedd233bcbbccd8c0bcd2fdd186216094d095f60076201538d"}, - {file = "pyzmq-26.4.0-cp313-cp313t-manylinux_2_28_x86_64.whl", hash = "sha256:237b283044934d26f1eeff4075f751b05d2f3ed42a257fc44386d00df6a270cf"}, - {file = "pyzmq-26.4.0-cp313-cp313t-musllinux_1_1_aarch64.whl", hash = "sha256:b30f862f6768b17040929a68432c8a8be77780317f45a353cb17e423127d250c"}, - {file = "pyzmq-26.4.0-cp313-cp313t-musllinux_1_1_i686.whl", hash = "sha256:c80fcd3504232f13617c6ab501124d373e4895424e65de8b72042333316f64a8"}, - {file = "pyzmq-26.4.0-cp313-cp313t-musllinux_1_1_x86_64.whl", hash = "sha256:26a2a7451606b87f67cdeca2c2789d86f605da08b4bd616b1a9981605ca3a364"}, - {file = "pyzmq-26.4.0-cp38-cp38-macosx_10_15_universal2.whl", hash = "sha256:831cc53bf6068d46d942af52fa8b0b9d128fb39bcf1f80d468dc9a3ae1da5bfb"}, - {file = "pyzmq-26.4.0-cp38-cp38-manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:51d18be6193c25bd229524cfac21e39887c8d5e0217b1857998dfbef57c070a4"}, - {file = "pyzmq-26.4.0-cp38-cp38-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:445c97854204119ae2232503585ebb4fa7517142f71092cb129e5ee547957a1f"}, - {file = "pyzmq-26.4.0-cp38-cp38-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:807b8f4ad3e6084412c0f3df0613269f552110fa6fb91743e3e306223dbf11a6"}, - {file = "pyzmq-26.4.0-cp38-cp38-musllinux_1_1_aarch64.whl", hash = "sha256:c01d109dd675ac47fa15c0a79d256878d898f90bc10589f808b62d021d2e653c"}, - {file = "pyzmq-26.4.0-cp38-cp38-musllinux_1_1_i686.whl", hash = "sha256:0a294026e28679a8dd64c922e59411cb586dad307661b4d8a5c49e7bbca37621"}, - {file = "pyzmq-26.4.0-cp38-cp38-musllinux_1_1_x86_64.whl", hash = "sha256:22c8dd677274af8dfb1efd05006d6f68fb2f054b17066e308ae20cb3f61028cf"}, - {file = "pyzmq-26.4.0-cp38-cp38-win32.whl", hash = "sha256:14fc678b696bc42c14e2d7f86ac4e97889d5e6b94d366ebcb637a768d2ad01af"}, - {file = "pyzmq-26.4.0-cp38-cp38-win_amd64.whl", hash = "sha256:d1ef0a536662bbbdc8525f7e2ef19e74123ec9c4578e0582ecd41aedc414a169"}, - {file = "pyzmq-26.4.0-cp39-cp39-macosx_10_15_universal2.whl", hash = "sha256:a88643de8abd000ce99ca72056a1a2ae15881ee365ecb24dd1d9111e43d57842"}, - {file = "pyzmq-26.4.0-cp39-cp39-manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:0a744ce209ecb557406fb928f3c8c55ce79b16c3eeb682da38ef5059a9af0848"}, - {file = "pyzmq-26.4.0-cp39-cp39-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:9434540f333332224ecb02ee6278b6c6f11ea1266b48526e73c903119b2f420f"}, - {file = "pyzmq-26.4.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e6c6f0a23e55cd38d27d4c89add963294ea091ebcb104d7fdab0f093bc5abb1c"}, - {file = "pyzmq-26.4.0-cp39-cp39-musllinux_1_1_aarch64.whl", hash = "sha256:6145df55dc2309f6ef72d70576dcd5aabb0fd373311613fe85a5e547c722b780"}, - {file = "pyzmq-26.4.0-cp39-cp39-musllinux_1_1_i686.whl", hash = "sha256:2ea81823840ef8c56e5d2f9918e4d571236294fea4d1842b302aebffb9e40997"}, - {file = "pyzmq-26.4.0-cp39-cp39-musllinux_1_1_x86_64.whl", hash = "sha256:cc2abc385dc37835445abe206524fbc0c9e3fce87631dfaa90918a1ba8f425eb"}, - {file = "pyzmq-26.4.0-cp39-cp39-win32.whl", hash = "sha256:41a2508fe7bed4c76b4cf55aacfb8733926f59d440d9ae2b81ee8220633b4d12"}, - {file = "pyzmq-26.4.0-cp39-cp39-win_amd64.whl", hash = "sha256:d4000e8255d6cbce38982e5622ebb90823f3409b7ffe8aeae4337ef7d6d2612a"}, - {file = "pyzmq-26.4.0-cp39-cp39-win_arm64.whl", hash = "sha256:b4f6919d9c120488246bdc2a2f96662fa80d67b35bd6d66218f457e722b3ff64"}, - {file = "pyzmq-26.4.0-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:98d948288ce893a2edc5ec3c438fe8de2daa5bbbd6e2e865ec5f966e237084ba"}, - {file = "pyzmq-26.4.0-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a9f34f5c9e0203ece706a1003f1492a56c06c0632d86cb77bcfe77b56aacf27b"}, - {file = "pyzmq-26.4.0-pp310-pypy310_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:80c9b48aef586ff8b698359ce22f9508937c799cc1d2c9c2f7c95996f2300c94"}, - {file = "pyzmq-26.4.0-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f3f2a5b74009fd50b53b26f65daff23e9853e79aa86e0aa08a53a7628d92d44a"}, - {file = "pyzmq-26.4.0-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:61c5f93d7622d84cb3092d7f6398ffc77654c346545313a3737e266fc11a3beb"}, - {file = "pyzmq-26.4.0-pp311-pypy311_pp73-macosx_10_15_x86_64.whl", hash = "sha256:4478b14cb54a805088299c25a79f27eaf530564a7a4f72bf432a040042b554eb"}, - {file = "pyzmq-26.4.0-pp311-pypy311_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8a28ac29c60e4ba84b5f58605ace8ad495414a724fe7aceb7cf06cd0598d04e1"}, - {file = "pyzmq-26.4.0-pp311-pypy311_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:43b03c1ceea27c6520124f4fb2ba9c647409b9abdf9a62388117148a90419494"}, - {file = "pyzmq-26.4.0-pp311-pypy311_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:7731abd23a782851426d4e37deb2057bf9410848a4459b5ede4fe89342e687a9"}, - {file = "pyzmq-26.4.0-pp311-pypy311_pp73-win_amd64.whl", hash = "sha256:a222ad02fbe80166b0526c038776e8042cd4e5f0dec1489a006a1df47e9040e0"}, - {file = "pyzmq-26.4.0-pp38-pypy38_pp73-macosx_10_15_x86_64.whl", hash = "sha256:91c3ffaea475ec8bb1a32d77ebc441dcdd13cd3c4c284a6672b92a0f5ade1917"}, - {file = "pyzmq-26.4.0-pp38-pypy38_pp73-manylinux_2_12_i686.manylinux2010_i686.whl", hash = "sha256:d9a78a52668bf5c9e7b0da36aa5760a9fc3680144e1445d68e98df78a25082ed"}, - {file = "pyzmq-26.4.0-pp38-pypy38_pp73-manylinux_2_12_x86_64.manylinux2010_x86_64.whl", hash = "sha256:b70cab356ff8c860118b89dc86cd910c73ce2127eb986dada4fbac399ef644cf"}, - {file = "pyzmq-26.4.0-pp38-pypy38_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:acae207d4387780838192326b32d373bb286da0b299e733860e96f80728eb0af"}, - {file = "pyzmq-26.4.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:f928eafd15794aa4be75463d537348b35503c1e014c5b663f206504ec1a90fe4"}, - {file = "pyzmq-26.4.0-pp39-pypy39_pp73-macosx_10_15_x86_64.whl", hash = "sha256:552b0d2e39987733e1e9e948a0ced6ff75e0ea39ab1a1db2fc36eb60fd8760db"}, - {file = "pyzmq-26.4.0-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:dd670a8aa843f2ee637039bbd412e0d7294a5e588e1ecc9ad98b0cdc050259a4"}, - {file = "pyzmq-26.4.0-pp39-pypy39_pp73-manylinux_2_17_i686.manylinux2014_i686.whl", hash = "sha256:d367b7b775a0e1e54a59a2ba3ed4d5e0a31566af97cc9154e34262777dab95ed"}, - {file = "pyzmq-26.4.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:8112af16c406e4a93df2caef49f884f4c2bb2b558b0b5577ef0b2465d15c1abc"}, - {file = "pyzmq-26.4.0-pp39-pypy39_pp73-manylinux_2_28_x86_64.whl", hash = "sha256:c76c298683f82669cab0b6da59071f55238c039738297c69f187a542c6d40099"}, - {file = "pyzmq-26.4.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:49b6ca2e625b46f499fb081aaf7819a177f41eeb555acb05758aa97f4f95d147"}, - {file = "pyzmq-26.4.0.tar.gz", hash = "sha256:4bd13f85f80962f91a651a7356fe0472791a5f7a92f227822b5acf44795c626d"}, + {file = "pyzmq-27.0.0-cp310-cp310-macosx_10_15_universal2.whl", hash = "sha256:b973ee650e8f442ce482c1d99ca7ab537c69098d53a3d046676a484fd710c87a"}, + {file = "pyzmq-27.0.0-cp310-cp310-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:661942bc7cd0223d569d808f2e5696d9cc120acc73bf3e88a1f1be7ab648a7e4"}, + {file = "pyzmq-27.0.0-cp310-cp310-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:50360fb2a056ffd16e5f4177eee67f1dd1017332ea53fb095fe7b5bf29c70246"}, + {file = "pyzmq-27.0.0-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:cf209a6dc4b420ed32a7093642843cbf8703ed0a7d86c16c0b98af46762ebefb"}, + {file = "pyzmq-27.0.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:c2dace4a7041cca2fba5357a2d7c97c5effdf52f63a1ef252cfa496875a3762d"}, + {file = "pyzmq-27.0.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:63af72b2955fc77caf0a77444baa2431fcabb4370219da38e1a9f8d12aaebe28"}, + {file = "pyzmq-27.0.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:e8c4adce8e37e75c4215297d7745551b8dcfa5f728f23ce09bf4e678a9399413"}, + {file = "pyzmq-27.0.0-cp310-cp310-win32.whl", hash = "sha256:5d5ef4718ecab24f785794e0e7536436698b459bfbc19a1650ef55280119d93b"}, + {file = "pyzmq-27.0.0-cp310-cp310-win_amd64.whl", hash = "sha256:e40609380480b3d12c30f841323f42451c755b8fece84235236f5fe5ffca8c1c"}, + {file = "pyzmq-27.0.0-cp310-cp310-win_arm64.whl", hash = "sha256:6b0397b0be277b46762956f576e04dc06ced265759e8c2ff41a0ee1aa0064198"}, + {file = "pyzmq-27.0.0-cp311-cp311-macosx_10_15_universal2.whl", hash = "sha256:21457825249b2a53834fa969c69713f8b5a79583689387a5e7aed880963ac564"}, + {file = "pyzmq-27.0.0-cp311-cp311-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:1958947983fef513e6e98eff9cb487b60bf14f588dc0e6bf35fa13751d2c8251"}, + {file = "pyzmq-27.0.0-cp311-cp311-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:c0dc628b5493f9a8cd9844b8bee9732ef587ab00002157c9329e4fc0ef4d3afa"}, + {file = "pyzmq-27.0.0-cp311-cp311-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:f7bbe9e1ed2c8d3da736a15694d87c12493e54cc9dc9790796f0321794bbc91f"}, + {file = "pyzmq-27.0.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:dc1091f59143b471d19eb64f54bae4f54bcf2a466ffb66fe45d94d8d734eb495"}, + {file = "pyzmq-27.0.0-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:7011ade88c8e535cf140f8d1a59428676fbbce7c6e54fefce58bf117aefb6667"}, + {file = "pyzmq-27.0.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:2c386339d7e3f064213aede5d03d054b237937fbca6dd2197ac8cf3b25a6b14e"}, + {file = "pyzmq-27.0.0-cp311-cp311-win32.whl", hash = "sha256:0546a720c1f407b2172cb04b6b094a78773491497e3644863cf5c96c42df8cff"}, + {file = "pyzmq-27.0.0-cp311-cp311-win_amd64.whl", hash = "sha256:15f39d50bd6c9091c67315ceb878a4f531957b121d2a05ebd077eb35ddc5efed"}, + {file = "pyzmq-27.0.0-cp311-cp311-win_arm64.whl", hash = "sha256:c5817641eebb391a2268c27fecd4162448e03538387093cdbd8bf3510c316b38"}, + {file = "pyzmq-27.0.0-cp312-abi3-macosx_10_15_universal2.whl", hash = "sha256:cbabc59dcfaac66655c040dfcb8118f133fb5dde185e5fc152628354c1598e52"}, + {file = "pyzmq-27.0.0-cp312-abi3-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:cb0ac5179cba4b2f94f1aa208fbb77b62c4c9bf24dd446278b8b602cf85fcda3"}, + {file = "pyzmq-27.0.0-cp312-abi3-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:53a48f0228eab6cbf69fde3aa3c03cbe04e50e623ef92ae395fce47ef8a76152"}, + {file = "pyzmq-27.0.0-cp312-abi3-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:111db5f395e09f7e775f759d598f43cb815fc58e0147623c4816486e1a39dc22"}, + {file = "pyzmq-27.0.0-cp312-abi3-musllinux_1_2_aarch64.whl", hash = "sha256:c8878011653dcdc27cc2c57e04ff96f0471e797f5c19ac3d7813a245bcb24371"}, + {file = "pyzmq-27.0.0-cp312-abi3-musllinux_1_2_i686.whl", hash = "sha256:c0ed2c1f335ba55b5fdc964622254917d6b782311c50e138863eda409fbb3b6d"}, + {file = "pyzmq-27.0.0-cp312-abi3-musllinux_1_2_x86_64.whl", hash = "sha256:e918d70862d4cfd4b1c187310015646a14e1f5917922ab45b29f28f345eeb6be"}, + {file = "pyzmq-27.0.0-cp312-abi3-win32.whl", hash = "sha256:88b4e43cab04c3c0f0d55df3b1eef62df2b629a1a369b5289a58f6fa8b07c4f4"}, + {file = "pyzmq-27.0.0-cp312-abi3-win_amd64.whl", hash = "sha256:dce4199bf5f648a902ce37e7b3afa286f305cd2ef7a8b6ec907470ccb6c8b371"}, + {file = "pyzmq-27.0.0-cp312-abi3-win_arm64.whl", hash = "sha256:56e46bbb85d52c1072b3f809cc1ce77251d560bc036d3a312b96db1afe76db2e"}, + {file = "pyzmq-27.0.0-cp313-cp313t-macosx_10_15_universal2.whl", hash = "sha256:c36ad534c0c29b4afa088dc53543c525b23c0797e01b69fef59b1a9c0e38b688"}, + {file = "pyzmq-27.0.0-cp313-cp313t-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:67855c14173aec36395d7777aaba3cc527b393821f30143fd20b98e1ff31fd38"}, + {file = "pyzmq-27.0.0-cp313-cp313t-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:8617c7d43cd8ccdb62aebe984bfed77ca8f036e6c3e46dd3dddda64b10f0ab7a"}, + {file = "pyzmq-27.0.0-cp313-cp313t-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:67bfbcbd0a04c575e8103a6061d03e393d9f80ffdb9beb3189261e9e9bc5d5e9"}, + {file = "pyzmq-27.0.0-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:5cd11d46d7b7e5958121b3eaf4cd8638eff3a720ec527692132f05a57f14341d"}, + {file = "pyzmq-27.0.0-cp313-cp313t-musllinux_1_2_i686.whl", hash = "sha256:b801c2e40c5aa6072c2f4876de8dccd100af6d9918d4d0d7aa54a1d982fd4f44"}, + {file = "pyzmq-27.0.0-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:20d5cb29e8c5f76a127c75b6e7a77e846bc4b655c373baa098c26a61b7ecd0ef"}, + {file = "pyzmq-27.0.0-cp313-cp313t-win32.whl", hash = "sha256:a20528da85c7ac7a19b7384e8c3f8fa707841fd85afc4ed56eda59d93e3d98ad"}, + {file = "pyzmq-27.0.0-cp313-cp313t-win_amd64.whl", hash = "sha256:d8229f2efece6a660ee211d74d91dbc2a76b95544d46c74c615e491900dc107f"}, + {file = "pyzmq-27.0.0-cp38-cp38-macosx_10_15_universal2.whl", hash = "sha256:f4162dbbd9c5c84fb930a36f290b08c93e35fce020d768a16fc8891a2f72bab8"}, + {file = "pyzmq-27.0.0-cp38-cp38-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:4e7d0a8d460fba526cc047333bdcbf172a159b8bd6be8c3eb63a416ff9ba1477"}, + {file = "pyzmq-27.0.0-cp38-cp38-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:29f44e3c26b9783816ba9ce274110435d8f5b19bbd82f7a6c7612bb1452a3597"}, + {file = "pyzmq-27.0.0-cp38-cp38-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:6e435540fa1da54667f0026cf1e8407fe6d8a11f1010b7f06b0b17214ebfcf5e"}, + {file = "pyzmq-27.0.0-cp38-cp38-musllinux_1_2_aarch64.whl", hash = "sha256:51f5726de3532b8222e569990c8aa34664faa97038304644679a51d906e60c6e"}, + {file = "pyzmq-27.0.0-cp38-cp38-musllinux_1_2_i686.whl", hash = "sha256:42c7555123679637c99205b1aa9e8f7d90fe29d4c243c719e347d4852545216c"}, + {file = "pyzmq-27.0.0-cp38-cp38-musllinux_1_2_x86_64.whl", hash = "sha256:a979b7cf9e33d86c4949df527a3018767e5f53bc3b02adf14d4d8db1db63ccc0"}, + {file = "pyzmq-27.0.0-cp38-cp38-win32.whl", hash = "sha256:26b72c5ae20bf59061c3570db835edb81d1e0706ff141747055591c4b41193f8"}, + {file = "pyzmq-27.0.0-cp38-cp38-win_amd64.whl", hash = "sha256:55a0155b148fe0428285a30922f7213539aa84329a5ad828bca4bbbc665c70a4"}, + {file = "pyzmq-27.0.0-cp39-cp39-macosx_10_15_universal2.whl", hash = "sha256:100f6e5052ba42b2533011d34a018a5ace34f8cac67cb03cfa37c8bdae0ca617"}, + {file = "pyzmq-27.0.0-cp39-cp39-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:bf6c6b061efd00404b9750e2cfbd9507492c8d4b3721ded76cb03786131be2ed"}, + {file = "pyzmq-27.0.0-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:ee05728c0b0b2484a9fc20466fa776fffb65d95f7317a3419985b8c908563861"}, + {file = "pyzmq-27.0.0-cp39-cp39-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:7cdf07fe0a557b131366f80727ec8ccc4b70d89f1e3f920d94a594d598d754f0"}, + {file = "pyzmq-27.0.0-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:90252fa2ff3a104219db1f5ced7032a7b5fc82d7c8d2fec2b9a3e6fd4e25576b"}, + {file = "pyzmq-27.0.0-cp39-cp39-musllinux_1_2_i686.whl", hash = "sha256:ea6d441c513bf18c578c73c323acf7b4184507fc244762193aa3a871333c9045"}, + {file = "pyzmq-27.0.0-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:ae2b34bcfaae20c064948a4113bf8709eee89fd08317eb293ae4ebd69b4d9740"}, + {file = "pyzmq-27.0.0-cp39-cp39-win32.whl", hash = "sha256:5b10bd6f008937705cf6e7bf8b6ece5ca055991e3eb130bca8023e20b86aa9a3"}, + {file = "pyzmq-27.0.0-cp39-cp39-win_amd64.whl", hash = "sha256:00387d12a8af4b24883895f7e6b9495dc20a66027b696536edac35cb988c38f3"}, + {file = "pyzmq-27.0.0-cp39-cp39-win_arm64.whl", hash = "sha256:4c19d39c04c29a6619adfeb19e3735c421b3bfee082f320662f52e59c47202ba"}, + {file = "pyzmq-27.0.0-pp310-pypy310_pp73-macosx_10_15_x86_64.whl", hash = "sha256:656c1866505a5735d0660b7da6d7147174bbf59d4975fc2b7f09f43c9bc25745"}, + {file = "pyzmq-27.0.0-pp310-pypy310_pp73-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:74175b9e12779382432dd1d1f5960ebe7465d36649b98a06c6b26be24d173fab"}, + {file = "pyzmq-27.0.0-pp310-pypy310_pp73-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:d8c6de908465697a8708e4d6843a1e884f567962fc61eb1706856545141d0cbb"}, + {file = "pyzmq-27.0.0-pp310-pypy310_pp73-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:c644aaacc01d0df5c7072826df45e67301f191c55f68d7b2916d83a9ddc1b551"}, + {file = "pyzmq-27.0.0-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:10f70c1d9a446a85013a36871a296007f6fe4232b530aa254baf9da3f8328bc0"}, + {file = "pyzmq-27.0.0-pp311-pypy311_pp73-macosx_10_15_x86_64.whl", hash = "sha256:cd1dc59763effd1576f8368047c9c31468fce0af89d76b5067641137506792ae"}, + {file = "pyzmq-27.0.0-pp311-pypy311_pp73-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:60e8cc82d968174650c1860d7b716366caab9973787a1c060cf8043130f7d0f7"}, + {file = "pyzmq-27.0.0-pp311-pypy311_pp73-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:14fe7aaac86e4e93ea779a821967360c781d7ac5115b3f1a171ced77065a0174"}, + {file = "pyzmq-27.0.0-pp311-pypy311_pp73-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:6ad0562d4e6abb785be3e4dd68599c41be821b521da38c402bc9ab2a8e7ebc7e"}, + {file = "pyzmq-27.0.0-pp311-pypy311_pp73-win_amd64.whl", hash = "sha256:9df43a2459cd3a3563404c1456b2c4c69564daa7dbaf15724c09821a3329ce46"}, + {file = "pyzmq-27.0.0-pp38-pypy38_pp73-macosx_10_15_x86_64.whl", hash = "sha256:8c86ea8fe85e2eb0ffa00b53192c401477d5252f6dd1db2e2ed21c1c30d17e5e"}, + {file = "pyzmq-27.0.0-pp38-pypy38_pp73-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:c45fee3968834cd291a13da5fac128b696c9592a9493a0f7ce0b47fa03cc574d"}, + {file = "pyzmq-27.0.0-pp38-pypy38_pp73-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:cae73bb6898c4e045fbed5024cb587e4110fddb66f6163bcab5f81f9d4b9c496"}, + {file = "pyzmq-27.0.0-pp38-pypy38_pp73-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:26d542258c7a1f35a9cff3d887687d3235006134b0ac1c62a6fe1ad3ac10440e"}, + {file = "pyzmq-27.0.0-pp38-pypy38_pp73-win_amd64.whl", hash = "sha256:04cd50ef3b28e35ced65740fb9956a5b3f77a6ff32fcd887e3210433f437dd0f"}, + {file = "pyzmq-27.0.0-pp39-pypy39_pp73-macosx_10_15_x86_64.whl", hash = "sha256:39ddd3ba0a641f01d8f13a3cfd4c4924eb58e660d8afe87e9061d6e8ca6f7ac3"}, + {file = "pyzmq-27.0.0-pp39-pypy39_pp73-manylinux2014_i686.manylinux_2_17_i686.whl", hash = "sha256:8ca7e6a0388dd9e1180b14728051068f4efe83e0d2de058b5ff92c63f399a73f"}, + {file = "pyzmq-27.0.0-pp39-pypy39_pp73-manylinux2014_x86_64.manylinux_2_17_x86_64.whl", hash = "sha256:2524c40891be6a3106885a3935d58452dd83eb7a5742a33cc780a1ad4c49dec0"}, + {file = "pyzmq-27.0.0-pp39-pypy39_pp73-manylinux_2_27_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:6a56e3e5bd2d62a01744fd2f1ce21d760c7c65f030e9522738d75932a14ab62a"}, + {file = "pyzmq-27.0.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:096af9e133fec3a72108ddefba1e42985cb3639e9de52cfd336b6fc23aa083e9"}, + {file = "pyzmq-27.0.0.tar.gz", hash = "sha256:b1f08eeb9ce1510e6939b6e5dcd46a17765e2333daae78ecf4606808442e52cf"}, ] [package.dependencies] @@ -3828,129 +3874,156 @@ jupyter = ["ipywidgets (>=7.5.1,<9)"] [[package]] name = "rpds-py" -version = "0.25.1" +version = "0.26.0" description = "Python bindings to Rust's persistent data structures (rpds)" optional = false python-versions = ">=3.9" groups = ["main", "test"] files = [ - {file = "rpds_py-0.25.1-cp310-cp310-macosx_10_12_x86_64.whl", hash = "sha256:f4ad628b5174d5315761b67f212774a32f5bad5e61396d38108bd801c0a8f5d9"}, - {file = "rpds_py-0.25.1-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:8c742af695f7525e559c16f1562cf2323db0e3f0fbdcabdf6865b095256b2d40"}, - {file = "rpds_py-0.25.1-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:605ffe7769e24b1800b4d024d24034405d9404f0bc2f55b6db3362cd34145a6f"}, - {file = "rpds_py-0.25.1-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:ccc6f3ddef93243538be76f8e47045b4aad7a66a212cd3a0f23e34469473d36b"}, - {file = "rpds_py-0.25.1-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:f70316f760174ca04492b5ab01be631a8ae30cadab1d1081035136ba12738cfa"}, - {file = "rpds_py-0.25.1-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e1dafef8df605fdb46edcc0bf1573dea0d6d7b01ba87f85cd04dc855b2b4479e"}, - {file = "rpds_py-0.25.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:0701942049095741a8aeb298a31b203e735d1c61f4423511d2b1a41dcd8a16da"}, - {file = "rpds_py-0.25.1-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:e87798852ae0b37c88babb7f7bbbb3e3fecc562a1c340195b44c7e24d403e380"}, - {file = "rpds_py-0.25.1-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:3bcce0edc1488906c2d4c75c94c70a0417e83920dd4c88fec1078c94843a6ce9"}, - {file = "rpds_py-0.25.1-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:e2f6a2347d3440ae789505693a02836383426249d5293541cd712e07e7aecf54"}, - {file = "rpds_py-0.25.1-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:4fd52d3455a0aa997734f3835cbc4c9f32571345143960e7d7ebfe7b5fbfa3b2"}, - {file = "rpds_py-0.25.1-cp310-cp310-win32.whl", hash = "sha256:3f0b1798cae2bbbc9b9db44ee068c556d4737911ad53a4e5093d09d04b3bbc24"}, - {file = "rpds_py-0.25.1-cp310-cp310-win_amd64.whl", hash = "sha256:3ebd879ab996537fc510a2be58c59915b5dd63bccb06d1ef514fee787e05984a"}, - {file = "rpds_py-0.25.1-cp311-cp311-macosx_10_12_x86_64.whl", hash = "sha256:5f048bbf18b1f9120685c6d6bb70cc1a52c8cc11bdd04e643d28d3be0baf666d"}, - {file = "rpds_py-0.25.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:4fbb0dbba559959fcb5d0735a0f87cdbca9e95dac87982e9b95c0f8f7ad10255"}, - {file = "rpds_py-0.25.1-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d4ca54b9cf9d80b4016a67a0193ebe0bcf29f6b0a96f09db942087e294d3d4c2"}, - {file = "rpds_py-0.25.1-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:1ee3e26eb83d39b886d2cb6e06ea701bba82ef30a0de044d34626ede51ec98b0"}, - {file = "rpds_py-0.25.1-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:89706d0683c73a26f76a5315d893c051324d771196ae8b13e6ffa1ffaf5e574f"}, - {file = "rpds_py-0.25.1-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c2013ee878c76269c7b557a9a9c042335d732e89d482606990b70a839635feb7"}, - {file = "rpds_py-0.25.1-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:45e484db65e5380804afbec784522de84fa95e6bb92ef1bd3325d33d13efaebd"}, - {file = "rpds_py-0.25.1-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:48d64155d02127c249695abb87d39f0faf410733428d499867606be138161d65"}, - {file = "rpds_py-0.25.1-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:048893e902132fd6548a2e661fb38bf4896a89eea95ac5816cf443524a85556f"}, - {file = "rpds_py-0.25.1-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:0317177b1e8691ab5879f4f33f4b6dc55ad3b344399e23df2e499de7b10a548d"}, - {file = "rpds_py-0.25.1-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:bffcf57826d77a4151962bf1701374e0fc87f536e56ec46f1abdd6a903354042"}, - {file = "rpds_py-0.25.1-cp311-cp311-win32.whl", hash = "sha256:cda776f1967cb304816173b30994faaf2fd5bcb37e73118a47964a02c348e1bc"}, - {file = "rpds_py-0.25.1-cp311-cp311-win_amd64.whl", hash = "sha256:dc3c1ff0abc91444cd20ec643d0f805df9a3661fcacf9c95000329f3ddf268a4"}, - {file = "rpds_py-0.25.1-cp311-cp311-win_arm64.whl", hash = "sha256:5a3ddb74b0985c4387719fc536faced33cadf2172769540c62e2a94b7b9be1c4"}, - {file = "rpds_py-0.25.1-cp312-cp312-macosx_10_12_x86_64.whl", hash = "sha256:b5ffe453cde61f73fea9430223c81d29e2fbf412a6073951102146c84e19e34c"}, - {file = "rpds_py-0.25.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:115874ae5e2fdcfc16b2aedc95b5eef4aebe91b28e7e21951eda8a5dc0d3461b"}, - {file = "rpds_py-0.25.1-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:a714bf6e5e81b0e570d01f56e0c89c6375101b8463999ead3a93a5d2a4af91fa"}, - {file = "rpds_py-0.25.1-cp312-cp312-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:35634369325906bcd01577da4c19e3b9541a15e99f31e91a02d010816b49bfda"}, - {file = "rpds_py-0.25.1-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:d4cb2b3ddc16710548801c6fcc0cfcdeeff9dafbc983f77265877793f2660309"}, - {file = "rpds_py-0.25.1-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9ceca1cf097ed77e1a51f1dbc8d174d10cb5931c188a4505ff9f3e119dfe519b"}, - {file = "rpds_py-0.25.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2c2cd1a4b0c2b8c5e31ffff50d09f39906fe351389ba143c195566056c13a7ea"}, - {file = "rpds_py-0.25.1-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:1de336a4b164c9188cb23f3703adb74a7623ab32d20090d0e9bf499a2203ad65"}, - {file = "rpds_py-0.25.1-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:9fca84a15333e925dd59ce01da0ffe2ffe0d6e5d29a9eeba2148916d1824948c"}, - {file = "rpds_py-0.25.1-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:88ec04afe0c59fa64e2f6ea0dd9657e04fc83e38de90f6de201954b4d4eb59bd"}, - {file = "rpds_py-0.25.1-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:a8bd2f19e312ce3e1d2c635618e8a8d8132892bb746a7cf74780a489f0f6cdcb"}, - {file = "rpds_py-0.25.1-cp312-cp312-win32.whl", hash = "sha256:e5e2f7280d8d0d3ef06f3ec1b4fd598d386cc6f0721e54f09109a8132182fbfe"}, - {file = "rpds_py-0.25.1-cp312-cp312-win_amd64.whl", hash = "sha256:db58483f71c5db67d643857404da360dce3573031586034b7d59f245144cc192"}, - {file = "rpds_py-0.25.1-cp312-cp312-win_arm64.whl", hash = "sha256:6d50841c425d16faf3206ddbba44c21aa3310a0cebc3c1cdfc3e3f4f9f6f5728"}, - {file = "rpds_py-0.25.1-cp313-cp313-macosx_10_12_x86_64.whl", hash = "sha256:659d87430a8c8c704d52d094f5ba6fa72ef13b4d385b7e542a08fc240cb4a559"}, - {file = "rpds_py-0.25.1-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:68f6f060f0bbdfb0245267da014d3a6da9be127fe3e8cc4a68c6f833f8a23bb1"}, - {file = "rpds_py-0.25.1-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:083a9513a33e0b92cf6e7a6366036c6bb43ea595332c1ab5c8ae329e4bcc0a9c"}, - {file = "rpds_py-0.25.1-cp313-cp313-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:816568614ecb22b18a010c7a12559c19f6fe993526af88e95a76d5a60b8b75fb"}, - {file = "rpds_py-0.25.1-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:3c6564c0947a7f52e4792983f8e6cf9bac140438ebf81f527a21d944f2fd0a40"}, - {file = "rpds_py-0.25.1-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:5c4a128527fe415d73cf1f70a9a688d06130d5810be69f3b553bf7b45e8acf79"}, - {file = "rpds_py-0.25.1-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:a49e1d7a4978ed554f095430b89ecc23f42014a50ac385eb0c4d163ce213c325"}, - {file = "rpds_py-0.25.1-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:d74ec9bc0e2feb81d3f16946b005748119c0f52a153f6db6a29e8cd68636f295"}, - {file = "rpds_py-0.25.1-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:3af5b4cc10fa41e5bc64e5c198a1b2d2864337f8fcbb9a67e747e34002ce812b"}, - {file = "rpds_py-0.25.1-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:79dc317a5f1c51fd9c6a0c4f48209c6b8526d0524a6904fc1076476e79b00f98"}, - {file = "rpds_py-0.25.1-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:1521031351865e0181bc585147624d66b3b00a84109b57fcb7a779c3ec3772cd"}, - {file = "rpds_py-0.25.1-cp313-cp313-win32.whl", hash = "sha256:5d473be2b13600b93a5675d78f59e63b51b1ba2d0476893415dfbb5477e65b31"}, - {file = "rpds_py-0.25.1-cp313-cp313-win_amd64.whl", hash = "sha256:a7b74e92a3b212390bdce1d93da9f6488c3878c1d434c5e751cbc202c5e09500"}, - {file = "rpds_py-0.25.1-cp313-cp313-win_arm64.whl", hash = "sha256:dd326a81afe332ede08eb39ab75b301d5676802cdffd3a8f287a5f0b694dc3f5"}, - {file = "rpds_py-0.25.1-cp313-cp313t-macosx_10_12_x86_64.whl", hash = "sha256:a58d1ed49a94d4183483a3ce0af22f20318d4a1434acee255d683ad90bf78129"}, - {file = "rpds_py-0.25.1-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:f251bf23deb8332823aef1da169d5d89fa84c89f67bdfb566c49dea1fccfd50d"}, - {file = "rpds_py-0.25.1-cp313-cp313t-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:8dbd586bfa270c1103ece2109314dd423df1fa3d9719928b5d09e4840cec0d72"}, - {file = "rpds_py-0.25.1-cp313-cp313t-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:6d273f136e912aa101a9274c3145dcbddbe4bac560e77e6d5b3c9f6e0ed06d34"}, - {file = "rpds_py-0.25.1-cp313-cp313t-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:666fa7b1bd0a3810a7f18f6d3a25ccd8866291fbbc3c9b912b917a6715874bb9"}, - {file = "rpds_py-0.25.1-cp313-cp313t-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:921954d7fbf3fccc7de8f717799304b14b6d9a45bbeec5a8d7408ccbf531faf5"}, - {file = "rpds_py-0.25.1-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:f3d86373ff19ca0441ebeb696ef64cb58b8b5cbacffcda5a0ec2f3911732a194"}, - {file = "rpds_py-0.25.1-cp313-cp313t-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:c8980cde3bb8575e7c956a530f2c217c1d6aac453474bf3ea0f9c89868b531b6"}, - {file = "rpds_py-0.25.1-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:8eb8c84ecea987a2523e057c0d950bcb3f789696c0499290b8d7b3107a719d78"}, - {file = "rpds_py-0.25.1-cp313-cp313t-musllinux_1_2_i686.whl", hash = "sha256:e43a005671a9ed5a650f3bc39e4dbccd6d4326b24fb5ea8be5f3a43a6f576c72"}, - {file = "rpds_py-0.25.1-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:58f77c60956501a4a627749a6dcb78dac522f249dd96b5c9f1c6af29bfacfb66"}, - {file = "rpds_py-0.25.1-cp313-cp313t-win32.whl", hash = "sha256:2cb9e5b5e26fc02c8a4345048cd9998c2aca7c2712bd1b36da0c72ee969a3523"}, - {file = "rpds_py-0.25.1-cp313-cp313t-win_amd64.whl", hash = "sha256:401ca1c4a20cc0510d3435d89c069fe0a9ae2ee6495135ac46bdd49ec0495763"}, - {file = "rpds_py-0.25.1-cp39-cp39-macosx_10_12_x86_64.whl", hash = "sha256:ce4c8e485a3c59593f1a6f683cf0ea5ab1c1dc94d11eea5619e4fb5228b40fbd"}, - {file = "rpds_py-0.25.1-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:d8222acdb51a22929c3b2ddb236b69c59c72af4019d2cba961e2f9add9b6e634"}, - {file = "rpds_py-0.25.1-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4593c4eae9b27d22df41cde518b4b9e4464d139e4322e2127daa9b5b981b76be"}, - {file = "rpds_py-0.25.1-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:bd035756830c712b64725a76327ce80e82ed12ebab361d3a1cdc0f51ea21acb0"}, - {file = "rpds_py-0.25.1-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:114a07e85f32b125404f28f2ed0ba431685151c037a26032b213c882f26eb908"}, - {file = "rpds_py-0.25.1-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:dec21e02e6cc932538b5203d3a8bd6aa1480c98c4914cb88eea064ecdbc6396a"}, - {file = "rpds_py-0.25.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:09eab132f41bf792c7a0ea1578e55df3f3e7f61888e340779b06050a9a3f16e9"}, - {file = "rpds_py-0.25.1-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:c98f126c4fc697b84c423e387337d5b07e4a61e9feac494362a59fd7a2d9ed80"}, - {file = "rpds_py-0.25.1-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:0e6a327af8ebf6baba1c10fadd04964c1965d375d318f4435d5f3f9651550f4a"}, - {file = "rpds_py-0.25.1-cp39-cp39-musllinux_1_2_i686.whl", hash = "sha256:bc120d1132cff853ff617754196d0ac0ae63befe7c8498bd67731ba368abe451"}, - {file = "rpds_py-0.25.1-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:140f61d9bed7839446bdd44852e30195c8e520f81329b4201ceead4d64eb3a9f"}, - {file = "rpds_py-0.25.1-cp39-cp39-win32.whl", hash = "sha256:9c006f3aadeda131b438c3092124bd196b66312f0caa5823ef09585a669cf449"}, - {file = "rpds_py-0.25.1-cp39-cp39-win_amd64.whl", hash = "sha256:a61d0b2c7c9a0ae45732a77844917b427ff16ad5464b4d4f5e4adb955f582890"}, - {file = "rpds_py-0.25.1-pp310-pypy310_pp73-macosx_10_12_x86_64.whl", hash = "sha256:b24bf3cd93d5b6ecfbedec73b15f143596c88ee249fa98cefa9a9dc9d92c6f28"}, - {file = "rpds_py-0.25.1-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:0eb90e94f43e5085623932b68840b6f379f26db7b5c2e6bcef3179bd83c9330f"}, - {file = "rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:d50e4864498a9ab639d6d8854b25e80642bd362ff104312d9770b05d66e5fb13"}, - {file = "rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:7c9409b47ba0650544b0bb3c188243b83654dfe55dcc173a86832314e1a6a35d"}, - {file = "rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:796ad874c89127c91970652a4ee8b00d56368b7e00d3477f4415fe78164c8000"}, - {file = "rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:85608eb70a659bf4c1142b2781083d4b7c0c4e2c90eff11856a9754e965b2540"}, - {file = "rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c4feb9211d15d9160bc85fa72fed46432cdc143eb9cf6d5ca377335a921ac37b"}, - {file = "rpds_py-0.25.1-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:ccfa689b9246c48947d31dd9d8b16d89a0ecc8e0e26ea5253068efb6c542b76e"}, - {file = "rpds_py-0.25.1-pp310-pypy310_pp73-musllinux_1_2_aarch64.whl", hash = "sha256:3c5b317ecbd8226887994852e85de562f7177add602514d4ac40f87de3ae45a8"}, - {file = "rpds_py-0.25.1-pp310-pypy310_pp73-musllinux_1_2_i686.whl", hash = "sha256:454601988aab2c6e8fd49e7634c65476b2b919647626208e376afcd22019eeb8"}, - {file = "rpds_py-0.25.1-pp310-pypy310_pp73-musllinux_1_2_x86_64.whl", hash = "sha256:1c0c434a53714358532d13539272db75a5ed9df75a4a090a753ac7173ec14e11"}, - {file = "rpds_py-0.25.1-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:f73ce1512e04fbe2bc97836e89830d6b4314c171587a99688082d090f934d20a"}, - {file = "rpds_py-0.25.1-pp311-pypy311_pp73-macosx_10_12_x86_64.whl", hash = "sha256:ee86d81551ec68a5c25373c5643d343150cc54672b5e9a0cafc93c1870a53954"}, - {file = "rpds_py-0.25.1-pp311-pypy311_pp73-macosx_11_0_arm64.whl", hash = "sha256:89c24300cd4a8e4a51e55c31a8ff3918e6651b241ee8876a42cc2b2a078533ba"}, - {file = "rpds_py-0.25.1-pp311-pypy311_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:771c16060ff4e79584dc48902a91ba79fd93eade3aa3a12d6d2a4aadaf7d542b"}, - {file = "rpds_py-0.25.1-pp311-pypy311_pp73-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:785ffacd0ee61c3e60bdfde93baa6d7c10d86f15655bd706c89da08068dc5038"}, - {file = "rpds_py-0.25.1-pp311-pypy311_pp73-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:2a40046a529cc15cef88ac5ab589f83f739e2d332cb4d7399072242400ed68c9"}, - {file = "rpds_py-0.25.1-pp311-pypy311_pp73-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:85fc223d9c76cabe5d0bff82214459189720dc135db45f9f66aa7cffbf9ff6c1"}, - {file = "rpds_py-0.25.1-pp311-pypy311_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:b0be9965f93c222fb9b4cc254235b3b2b215796c03ef5ee64f995b1b69af0762"}, - {file = "rpds_py-0.25.1-pp311-pypy311_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:8378fa4a940f3fb509c081e06cb7f7f2adae8cf46ef258b0e0ed7519facd573e"}, - {file = "rpds_py-0.25.1-pp311-pypy311_pp73-musllinux_1_2_aarch64.whl", hash = "sha256:33358883a4490287e67a2c391dfaea4d9359860281db3292b6886bf0be3d8692"}, - {file = "rpds_py-0.25.1-pp311-pypy311_pp73-musllinux_1_2_i686.whl", hash = "sha256:1d1fadd539298e70cac2f2cb36f5b8a65f742b9b9f1014dd4ea1f7785e2470bf"}, - {file = "rpds_py-0.25.1-pp311-pypy311_pp73-musllinux_1_2_x86_64.whl", hash = "sha256:9a46c2fb2545e21181445515960006e85d22025bd2fe6db23e76daec6eb689fe"}, - {file = "rpds_py-0.25.1-pp39-pypy39_pp73-macosx_10_12_x86_64.whl", hash = "sha256:50f2c501a89c9a5f4e454b126193c5495b9fb441a75b298c60591d8a2eb92e1b"}, - {file = "rpds_py-0.25.1-pp39-pypy39_pp73-macosx_11_0_arm64.whl", hash = "sha256:7d779b325cc8238227c47fbc53964c8cc9a941d5dbae87aa007a1f08f2f77b23"}, - {file = "rpds_py-0.25.1-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:036ded36bedb727beeabc16dc1dad7cb154b3fa444e936a03b67a86dc6a5066e"}, - {file = "rpds_py-0.25.1-pp39-pypy39_pp73-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:245550f5a1ac98504147cba96ffec8fabc22b610742e9150138e5d60774686d7"}, - {file = "rpds_py-0.25.1-pp39-pypy39_pp73-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:ff7c23ba0a88cb7b104281a99476cccadf29de2a0ef5ce864959a52675b1ca83"}, - {file = "rpds_py-0.25.1-pp39-pypy39_pp73-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:e37caa8cdb3b7cf24786451a0bdb853f6347b8b92005eeb64225ae1db54d1c2b"}, - {file = "rpds_py-0.25.1-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:9f2f48ab00181600ee266a095fe815134eb456163f7d6699f525dee471f312cf"}, - {file = "rpds_py-0.25.1-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:9e5fc7484fa7dce57e25063b0ec9638ff02a908304f861d81ea49273e43838c1"}, - {file = "rpds_py-0.25.1-pp39-pypy39_pp73-musllinux_1_2_aarch64.whl", hash = "sha256:d3c10228d6cf6fe2b63d2e7985e94f6916fa46940df46b70449e9ff9297bd3d1"}, - {file = "rpds_py-0.25.1-pp39-pypy39_pp73-musllinux_1_2_i686.whl", hash = "sha256:5d9e40f32745db28c1ef7aad23f6fc458dc1e29945bd6781060f0d15628b8ddf"}, - {file = "rpds_py-0.25.1-pp39-pypy39_pp73-musllinux_1_2_x86_64.whl", hash = "sha256:35a8d1a24b5936b35c5003313bc177403d8bdef0f8b24f28b1c4a255f94ea992"}, - {file = "rpds_py-0.25.1-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:6099263f526efff9cf3883dfef505518730f7a7a93049b1d90d42e50a22b4793"}, - {file = "rpds_py-0.25.1.tar.gz", hash = "sha256:8960b6dac09b62dac26e75d7e2c4a22efb835d827a7278c34f72b2b84fa160e3"}, + {file = "rpds_py-0.26.0-cp310-cp310-macosx_10_12_x86_64.whl", hash = "sha256:4c70c70f9169692b36307a95f3d8c0a9fcd79f7b4a383aad5eaa0e9718b79b37"}, + {file = "rpds_py-0.26.0-cp310-cp310-macosx_11_0_arm64.whl", hash = "sha256:777c62479d12395bfb932944e61e915741e364c843afc3196b694db3d669fcd0"}, + {file = "rpds_py-0.26.0-cp310-cp310-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:ec671691e72dff75817386aa02d81e708b5a7ec0dec6669ec05213ff6b77e1bd"}, + {file = "rpds_py-0.26.0-cp310-cp310-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:6a1cb5d6ce81379401bbb7f6dbe3d56de537fb8235979843f0d53bc2e9815a79"}, + {file = "rpds_py-0.26.0-cp310-cp310-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4f789e32fa1fb6a7bf890e0124e7b42d1e60d28ebff57fe806719abb75f0e9a3"}, + {file = "rpds_py-0.26.0-cp310-cp310-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9c55b0a669976cf258afd718de3d9ad1b7d1fe0a91cd1ab36f38b03d4d4aeaaf"}, + {file = "rpds_py-0.26.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c70d9ec912802ecfd6cd390dadb34a9578b04f9bcb8e863d0a7598ba5e9e7ccc"}, + {file = "rpds_py-0.26.0-cp310-cp310-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:3021933c2cb7def39d927b9862292e0f4c75a13d7de70eb0ab06efed4c508c19"}, + {file = "rpds_py-0.26.0-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:8a7898b6ca3b7d6659e55cdac825a2e58c638cbf335cde41f4619e290dd0ad11"}, + {file = "rpds_py-0.26.0-cp310-cp310-musllinux_1_2_i686.whl", hash = "sha256:12bff2ad9447188377f1b2794772f91fe68bb4bbfa5a39d7941fbebdbf8c500f"}, + {file = "rpds_py-0.26.0-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:191aa858f7d4902e975d4cf2f2d9243816c91e9605070aeb09c0a800d187e323"}, + {file = "rpds_py-0.26.0-cp310-cp310-win32.whl", hash = "sha256:b37a04d9f52cb76b6b78f35109b513f6519efb481d8ca4c321f6a3b9580b3f45"}, + {file = "rpds_py-0.26.0-cp310-cp310-win_amd64.whl", hash = "sha256:38721d4c9edd3eb6670437d8d5e2070063f305bfa2d5aa4278c51cedcd508a84"}, + {file = "rpds_py-0.26.0-cp311-cp311-macosx_10_12_x86_64.whl", hash = "sha256:9e8cb77286025bdb21be2941d64ac6ca016130bfdcd228739e8ab137eb4406ed"}, + {file = "rpds_py-0.26.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5e09330b21d98adc8ccb2dbb9fc6cb434e8908d4c119aeaa772cb1caab5440a0"}, + {file = "rpds_py-0.26.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:2c9c1b92b774b2e68d11193dc39620d62fd8ab33f0a3c77ecdabe19c179cdbc1"}, + {file = "rpds_py-0.26.0-cp311-cp311-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:824e6d3503ab990d7090768e4dfd9e840837bae057f212ff9f4f05ec6d1975e7"}, + {file = "rpds_py-0.26.0-cp311-cp311-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:8ad7fd2258228bf288f2331f0a6148ad0186b2e3643055ed0db30990e59817a6"}, + {file = "rpds_py-0.26.0-cp311-cp311-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:0dc23bbb3e06ec1ea72d515fb572c1fea59695aefbffb106501138762e1e915e"}, + {file = "rpds_py-0.26.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d80bf832ac7b1920ee29a426cdca335f96a2b5caa839811803e999b41ba9030d"}, + {file = "rpds_py-0.26.0-cp311-cp311-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:0919f38f5542c0a87e7b4afcafab6fd2c15386632d249e9a087498571250abe3"}, + {file = "rpds_py-0.26.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:d422b945683e409000c888e384546dbab9009bb92f7c0b456e217988cf316107"}, + {file = "rpds_py-0.26.0-cp311-cp311-musllinux_1_2_i686.whl", hash = "sha256:77a7711fa562ba2da1aa757e11024ad6d93bad6ad7ede5afb9af144623e5f76a"}, + {file = "rpds_py-0.26.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:238e8c8610cb7c29460e37184f6799547f7e09e6a9bdbdab4e8edb90986a2318"}, + {file = "rpds_py-0.26.0-cp311-cp311-win32.whl", hash = "sha256:893b022bfbdf26d7bedb083efeea624e8550ca6eb98bf7fea30211ce95b9201a"}, + {file = "rpds_py-0.26.0-cp311-cp311-win_amd64.whl", hash = "sha256:87a5531de9f71aceb8af041d72fc4cab4943648d91875ed56d2e629bef6d4c03"}, + {file = "rpds_py-0.26.0-cp311-cp311-win_arm64.whl", hash = "sha256:de2713f48c1ad57f89ac25b3cb7daed2156d8e822cf0eca9b96a6f990718cc41"}, + {file = "rpds_py-0.26.0-cp312-cp312-macosx_10_12_x86_64.whl", hash = "sha256:894514d47e012e794f1350f076c427d2347ebf82f9b958d554d12819849a369d"}, + {file = "rpds_py-0.26.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:fc921b96fa95a097add244da36a1d9e4f3039160d1d30f1b35837bf108c21136"}, + {file = "rpds_py-0.26.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:3e1157659470aa42a75448b6e943c895be8c70531c43cb78b9ba990778955582"}, + {file = "rpds_py-0.26.0-cp312-cp312-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:521ccf56f45bb3a791182dc6b88ae5f8fa079dd705ee42138c76deb1238e554e"}, + {file = "rpds_py-0.26.0-cp312-cp312-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9def736773fd56b305c0eef698be5192c77bfa30d55a0e5885f80126c4831a15"}, + {file = "rpds_py-0.26.0-cp312-cp312-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:cdad4ea3b4513b475e027be79e5a0ceac8ee1c113a1a11e5edc3c30c29f964d8"}, + {file = "rpds_py-0.26.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:82b165b07f416bdccf5c84546a484cc8f15137ca38325403864bfdf2b5b72f6a"}, + {file = "rpds_py-0.26.0-cp312-cp312-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:d04cab0a54b9dba4d278fe955a1390da3cf71f57feb78ddc7cb67cbe0bd30323"}, + {file = "rpds_py-0.26.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:79061ba1a11b6a12743a2b0f72a46aa2758613d454aa6ba4f5a265cc48850158"}, + {file = "rpds_py-0.26.0-cp312-cp312-musllinux_1_2_i686.whl", hash = "sha256:f405c93675d8d4c5ac87364bb38d06c988e11028a64b52a47158a355079661f3"}, + {file = "rpds_py-0.26.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:dafd4c44b74aa4bed4b250f1aed165b8ef5de743bcca3b88fc9619b6087093d2"}, + {file = "rpds_py-0.26.0-cp312-cp312-win32.whl", hash = "sha256:3da5852aad63fa0c6f836f3359647870e21ea96cf433eb393ffa45263a170d44"}, + {file = "rpds_py-0.26.0-cp312-cp312-win_amd64.whl", hash = "sha256:cf47cfdabc2194a669dcf7a8dbba62e37a04c5041d2125fae0233b720da6f05c"}, + {file = "rpds_py-0.26.0-cp312-cp312-win_arm64.whl", hash = "sha256:20ab1ae4fa534f73647aad289003f1104092890849e0266271351922ed5574f8"}, + {file = "rpds_py-0.26.0-cp313-cp313-macosx_10_12_x86_64.whl", hash = "sha256:696764a5be111b036256c0b18cd29783fab22154690fc698062fc1b0084b511d"}, + {file = "rpds_py-0.26.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:1e6c15d2080a63aaed876e228efe4f814bc7889c63b1e112ad46fdc8b368b9e1"}, + {file = "rpds_py-0.26.0-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:390e3170babf42462739a93321e657444f0862c6d722a291accc46f9d21ed04e"}, + {file = "rpds_py-0.26.0-cp313-cp313-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:7da84c2c74c0f5bc97d853d9e17bb83e2dcafcff0dc48286916001cc114379a1"}, + {file = "rpds_py-0.26.0-cp313-cp313-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:4c5fe114a6dd480a510b6d3661d09d67d1622c4bf20660a474507aaee7eeeee9"}, + {file = "rpds_py-0.26.0-cp313-cp313-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:3100b3090269f3a7ea727b06a6080d4eb7439dca4c0e91a07c5d133bb1727ea7"}, + {file = "rpds_py-0.26.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:2c03c9b0c64afd0320ae57de4c982801271c0c211aa2d37f3003ff5feb75bb04"}, + {file = "rpds_py-0.26.0-cp313-cp313-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:5963b72ccd199ade6ee493723d18a3f21ba7d5b957017607f815788cef50eaf1"}, + {file = "rpds_py-0.26.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:9da4e873860ad5bab3291438525cae80169daecbfafe5657f7f5fb4d6b3f96b9"}, + {file = "rpds_py-0.26.0-cp313-cp313-musllinux_1_2_i686.whl", hash = "sha256:5afaddaa8e8c7f1f7b4c5c725c0070b6eed0228f705b90a1732a48e84350f4e9"}, + {file = "rpds_py-0.26.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:4916dc96489616a6f9667e7526af8fa693c0fdb4f3acb0e5d9f4400eb06a47ba"}, + {file = "rpds_py-0.26.0-cp313-cp313-win32.whl", hash = "sha256:2a343f91b17097c546b93f7999976fd6c9d5900617aa848c81d794e062ab302b"}, + {file = "rpds_py-0.26.0-cp313-cp313-win_amd64.whl", hash = "sha256:0a0b60701f2300c81b2ac88a5fb893ccfa408e1c4a555a77f908a2596eb875a5"}, + {file = "rpds_py-0.26.0-cp313-cp313-win_arm64.whl", hash = "sha256:257d011919f133a4746958257f2c75238e3ff54255acd5e3e11f3ff41fd14256"}, + {file = "rpds_py-0.26.0-cp313-cp313t-macosx_10_12_x86_64.whl", hash = "sha256:529c8156d7506fba5740e05da8795688f87119cce330c244519cf706a4a3d618"}, + {file = "rpds_py-0.26.0-cp313-cp313t-macosx_11_0_arm64.whl", hash = "sha256:f53ec51f9d24e9638a40cabb95078ade8c99251945dad8d57bf4aabe86ecee35"}, + {file = "rpds_py-0.26.0-cp313-cp313t-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:7ab504c4d654e4a29558eaa5bb8cea5fdc1703ea60a8099ffd9c758472cf913f"}, + {file = "rpds_py-0.26.0-cp313-cp313t-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:fd0641abca296bc1a00183fe44f7fced8807ed49d501f188faa642d0e4975b83"}, + {file = "rpds_py-0.26.0-cp313-cp313t-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:69b312fecc1d017b5327afa81d4da1480f51c68810963a7336d92203dbb3d4f1"}, + {file = "rpds_py-0.26.0-cp313-cp313t-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:c741107203954f6fc34d3066d213d0a0c40f7bb5aafd698fb39888af277c70d8"}, + {file = "rpds_py-0.26.0-cp313-cp313t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:fc3e55a7db08dc9a6ed5fb7103019d2c1a38a349ac41901f9f66d7f95750942f"}, + {file = "rpds_py-0.26.0-cp313-cp313t-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:9e851920caab2dbcae311fd28f4313c6953993893eb5c1bb367ec69d9a39e7ed"}, + {file = "rpds_py-0.26.0-cp313-cp313t-musllinux_1_2_aarch64.whl", hash = "sha256:dfbf280da5f876d0b00c81f26bedce274e72a678c28845453885a9b3c22ae632"}, + {file = "rpds_py-0.26.0-cp313-cp313t-musllinux_1_2_i686.whl", hash = "sha256:1cc81d14ddfa53d7f3906694d35d54d9d3f850ef8e4e99ee68bc0d1e5fed9a9c"}, + {file = "rpds_py-0.26.0-cp313-cp313t-musllinux_1_2_x86_64.whl", hash = "sha256:dca83c498b4650a91efcf7b88d669b170256bf8017a5db6f3e06c2bf031f57e0"}, + {file = "rpds_py-0.26.0-cp313-cp313t-win32.whl", hash = "sha256:4d11382bcaf12f80b51d790dee295c56a159633a8e81e6323b16e55d81ae37e9"}, + {file = "rpds_py-0.26.0-cp313-cp313t-win_amd64.whl", hash = "sha256:ff110acded3c22c033e637dd8896e411c7d3a11289b2edf041f86663dbc791e9"}, + {file = "rpds_py-0.26.0-cp314-cp314-macosx_10_12_x86_64.whl", hash = "sha256:da619979df60a940cd434084355c514c25cf8eb4cf9a508510682f6c851a4f7a"}, + {file = "rpds_py-0.26.0-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:ea89a2458a1a75f87caabefe789c87539ea4e43b40f18cff526052e35bbb4fdf"}, + {file = "rpds_py-0.26.0-cp314-cp314-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:feac1045b3327a45944e7dcbeb57530339f6b17baff154df51ef8b0da34c8c12"}, + {file = "rpds_py-0.26.0-cp314-cp314-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:b818a592bd69bfe437ee8368603d4a2d928c34cffcdf77c2e761a759ffd17d20"}, + {file = "rpds_py-0.26.0-cp314-cp314-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:1a8b0dd8648709b62d9372fc00a57466f5fdeefed666afe3fea5a6c9539a0331"}, + {file = "rpds_py-0.26.0-cp314-cp314-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:6d3498ad0df07d81112aa6ec6c95a7e7b1ae00929fb73e7ebee0f3faaeabad2f"}, + {file = "rpds_py-0.26.0-cp314-cp314-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:24a4146ccb15be237fdef10f331c568e1b0e505f8c8c9ed5d67759dac58ac246"}, + {file = "rpds_py-0.26.0-cp314-cp314-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:a9a63785467b2d73635957d32a4f6e73d5e4df497a16a6392fa066b753e87387"}, + {file = "rpds_py-0.26.0-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:de4ed93a8c91debfd5a047be327b7cc8b0cc6afe32a716bbbc4aedca9e2a83af"}, + {file = "rpds_py-0.26.0-cp314-cp314-musllinux_1_2_i686.whl", hash = "sha256:caf51943715b12af827696ec395bfa68f090a4c1a1d2509eb4e2cb69abbbdb33"}, + {file = "rpds_py-0.26.0-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:4a59e5bc386de021f56337f757301b337d7ab58baa40174fb150accd480bc953"}, + {file = "rpds_py-0.26.0-cp314-cp314-win32.whl", hash = "sha256:92c8db839367ef16a662478f0a2fe13e15f2227da3c1430a782ad0f6ee009ec9"}, + {file = "rpds_py-0.26.0-cp314-cp314-win_amd64.whl", hash = "sha256:b0afb8cdd034150d4d9f53926226ed27ad15b7f465e93d7468caaf5eafae0d37"}, + {file = "rpds_py-0.26.0-cp314-cp314-win_arm64.whl", hash = "sha256:ca3f059f4ba485d90c8dc75cb5ca897e15325e4e609812ce57f896607c1c0867"}, + {file = "rpds_py-0.26.0-cp314-cp314t-macosx_10_12_x86_64.whl", hash = "sha256:5afea17ab3a126006dc2f293b14ffc7ef3c85336cf451564a0515ed7648033da"}, + {file = "rpds_py-0.26.0-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:69f0c0a3df7fd3a7eec50a00396104bb9a843ea6d45fcc31c2d5243446ffd7a7"}, + {file = "rpds_py-0.26.0-cp314-cp314t-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:801a71f70f9813e82d2513c9a96532551fce1e278ec0c64610992c49c04c2dad"}, + {file = "rpds_py-0.26.0-cp314-cp314t-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:df52098cde6d5e02fa75c1f6244f07971773adb4a26625edd5c18fee906fa84d"}, + {file = "rpds_py-0.26.0-cp314-cp314t-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:9bc596b30f86dc6f0929499c9e574601679d0341a0108c25b9b358a042f51bca"}, + {file = "rpds_py-0.26.0-cp314-cp314t-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:9dfbe56b299cf5875b68eb6f0ebaadc9cac520a1989cac0db0765abfb3709c19"}, + {file = "rpds_py-0.26.0-cp314-cp314t-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:ac64f4b2bdb4ea622175c9ab7cf09444e412e22c0e02e906978b3b488af5fde8"}, + {file = "rpds_py-0.26.0-cp314-cp314t-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:181ef9b6bbf9845a264f9aa45c31836e9f3c1f13be565d0d010e964c661d1e2b"}, + {file = "rpds_py-0.26.0-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:49028aa684c144ea502a8e847d23aed5e4c2ef7cadfa7d5eaafcb40864844b7a"}, + {file = "rpds_py-0.26.0-cp314-cp314t-musllinux_1_2_i686.whl", hash = "sha256:e5d524d68a474a9688336045bbf76cb0def88549c1b2ad9dbfec1fb7cfbe9170"}, + {file = "rpds_py-0.26.0-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:c1851f429b822831bd2edcbe0cfd12ee9ea77868f8d3daf267b189371671c80e"}, + {file = "rpds_py-0.26.0-cp314-cp314t-win32.whl", hash = "sha256:7bdb17009696214c3b66bb3590c6d62e14ac5935e53e929bcdbc5a495987a84f"}, + {file = "rpds_py-0.26.0-cp314-cp314t-win_amd64.whl", hash = "sha256:f14440b9573a6f76b4ee4770c13f0b5921f71dde3b6fcb8dabbefd13b7fe05d7"}, + {file = "rpds_py-0.26.0-cp39-cp39-macosx_10_12_x86_64.whl", hash = "sha256:7a48af25d9b3c15684059d0d1fc0bc30e8eee5ca521030e2bffddcab5be40226"}, + {file = "rpds_py-0.26.0-cp39-cp39-macosx_11_0_arm64.whl", hash = "sha256:0c71c2f6bf36e61ee5c47b2b9b5d47e4d1baad6426bfed9eea3e858fc6ee8806"}, + {file = "rpds_py-0.26.0-cp39-cp39-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:1d815d48b1804ed7867b539236b6dd62997850ca1c91cad187f2ddb1b7bbef19"}, + {file = "rpds_py-0.26.0-cp39-cp39-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:84cfbd4d4d2cdeb2be61a057a258d26b22877266dd905809e94172dff01a42ae"}, + {file = "rpds_py-0.26.0-cp39-cp39-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:fbaa70553ca116c77717f513e08815aec458e6b69a028d4028d403b3bc84ff37"}, + {file = "rpds_py-0.26.0-cp39-cp39-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:39bfea47c375f379d8e87ab4bb9eb2c836e4f2069f0f65731d85e55d74666387"}, + {file = "rpds_py-0.26.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1533b7eb683fb5f38c1d68a3c78f5fdd8f1412fa6b9bf03b40f450785a0ab915"}, + {file = "rpds_py-0.26.0-cp39-cp39-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:c5ab0ee51f560d179b057555b4f601b7df909ed31312d301b99f8b9fc6028284"}, + {file = "rpds_py-0.26.0-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:e5162afc9e0d1f9cae3b577d9c29ddbab3505ab39012cb794d94a005825bde21"}, + {file = "rpds_py-0.26.0-cp39-cp39-musllinux_1_2_i686.whl", hash = "sha256:43f10b007033f359bc3fa9cd5e6c1e76723f056ffa9a6b5c117cc35720a80292"}, + {file = "rpds_py-0.26.0-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:e3730a48e5622e598293eee0762b09cff34dd3f271530f47b0894891281f051d"}, + {file = "rpds_py-0.26.0-cp39-cp39-win32.whl", hash = "sha256:4b1f66eb81eab2e0ff5775a3a312e5e2e16bf758f7b06be82fb0d04078c7ac51"}, + {file = "rpds_py-0.26.0-cp39-cp39-win_amd64.whl", hash = "sha256:519067e29f67b5c90e64fb1a6b6e9d2ec0ba28705c51956637bac23a2f4ddae1"}, + {file = "rpds_py-0.26.0-pp310-pypy310_pp73-macosx_10_12_x86_64.whl", hash = "sha256:3c0909c5234543ada2515c05dc08595b08d621ba919629e94427e8e03539c958"}, + {file = "rpds_py-0.26.0-pp310-pypy310_pp73-macosx_11_0_arm64.whl", hash = "sha256:c1fb0cda2abcc0ac62f64e2ea4b4e64c57dfd6b885e693095460c61bde7bb18e"}, + {file = "rpds_py-0.26.0-pp310-pypy310_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:84d142d2d6cf9b31c12aa4878d82ed3b2324226270b89b676ac62ccd7df52d08"}, + {file = "rpds_py-0.26.0-pp310-pypy310_pp73-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:a547e21c5610b7e9093d870be50682a6a6cf180d6da0f42c47c306073bfdbbf6"}, + {file = "rpds_py-0.26.0-pp310-pypy310_pp73-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:35e9a70a0f335371275cdcd08bc5b8051ac494dd58bff3bbfb421038220dc871"}, + {file = "rpds_py-0.26.0-pp310-pypy310_pp73-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:0dfa6115c6def37905344d56fb54c03afc49104e2ca473d5dedec0f6606913b4"}, + {file = "rpds_py-0.26.0-pp310-pypy310_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:313cfcd6af1a55a286a3c9a25f64af6d0e46cf60bc5798f1db152d97a216ff6f"}, + {file = "rpds_py-0.26.0-pp310-pypy310_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:f7bf2496fa563c046d05e4d232d7b7fd61346e2402052064b773e5c378bf6f73"}, + {file = "rpds_py-0.26.0-pp310-pypy310_pp73-musllinux_1_2_aarch64.whl", hash = "sha256:aa81873e2c8c5aa616ab8e017a481a96742fdf9313c40f14338ca7dbf50cb55f"}, + {file = "rpds_py-0.26.0-pp310-pypy310_pp73-musllinux_1_2_i686.whl", hash = "sha256:68ffcf982715f5b5b7686bdd349ff75d422e8f22551000c24b30eaa1b7f7ae84"}, + {file = "rpds_py-0.26.0-pp310-pypy310_pp73-musllinux_1_2_x86_64.whl", hash = "sha256:6188de70e190847bb6db3dc3981cbadff87d27d6fe9b4f0e18726d55795cee9b"}, + {file = "rpds_py-0.26.0-pp310-pypy310_pp73-win_amd64.whl", hash = "sha256:1c962145c7473723df9722ba4c058de12eb5ebedcb4e27e7d902920aa3831ee8"}, + {file = "rpds_py-0.26.0-pp311-pypy311_pp73-macosx_10_12_x86_64.whl", hash = "sha256:f61a9326f80ca59214d1cceb0a09bb2ece5b2563d4e0cd37bfd5515c28510674"}, + {file = "rpds_py-0.26.0-pp311-pypy311_pp73-macosx_11_0_arm64.whl", hash = "sha256:183f857a53bcf4b1b42ef0f57ca553ab56bdd170e49d8091e96c51c3d69ca696"}, + {file = "rpds_py-0.26.0-pp311-pypy311_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:941c1cfdf4799d623cf3aa1d326a6b4fdb7a5799ee2687f3516738216d2262fb"}, + {file = "rpds_py-0.26.0-pp311-pypy311_pp73-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:72a8d9564a717ee291f554eeb4bfeafe2309d5ec0aa6c475170bdab0f9ee8e88"}, + {file = "rpds_py-0.26.0-pp311-pypy311_pp73-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:511d15193cbe013619dd05414c35a7dedf2088fcee93c6bbb7c77859765bd4e8"}, + {file = "rpds_py-0.26.0-pp311-pypy311_pp73-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:aea1f9741b603a8d8fedb0ed5502c2bc0accbc51f43e2ad1337fe7259c2b77a5"}, + {file = "rpds_py-0.26.0-pp311-pypy311_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4019a9d473c708cf2f16415688ef0b4639e07abaa569d72f74745bbeffafa2c7"}, + {file = "rpds_py-0.26.0-pp311-pypy311_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:093d63b4b0f52d98ebae33b8c50900d3d67e0666094b1be7a12fffd7f65de74b"}, + {file = "rpds_py-0.26.0-pp311-pypy311_pp73-musllinux_1_2_aarch64.whl", hash = "sha256:2abe21d8ba64cded53a2a677e149ceb76dcf44284202d737178afe7ba540c1eb"}, + {file = "rpds_py-0.26.0-pp311-pypy311_pp73-musllinux_1_2_i686.whl", hash = "sha256:4feb7511c29f8442cbbc28149a92093d32e815a28aa2c50d333826ad2a20fdf0"}, + {file = "rpds_py-0.26.0-pp311-pypy311_pp73-musllinux_1_2_x86_64.whl", hash = "sha256:e99685fc95d386da368013e7fb4269dd39c30d99f812a8372d62f244f662709c"}, + {file = "rpds_py-0.26.0-pp39-pypy39_pp73-macosx_10_12_x86_64.whl", hash = "sha256:a90a13408a7a856b87be8a9f008fff53c5080eea4e4180f6c2e546e4a972fb5d"}, + {file = "rpds_py-0.26.0-pp39-pypy39_pp73-macosx_11_0_arm64.whl", hash = "sha256:3ac51b65e8dc76cf4949419c54c5528adb24fc721df722fd452e5fbc236f5c40"}, + {file = "rpds_py-0.26.0-pp39-pypy39_pp73-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:59b2093224a18c6508d95cfdeba8db9cbfd6f3494e94793b58972933fcee4c6d"}, + {file = "rpds_py-0.26.0-pp39-pypy39_pp73-manylinux_2_17_armv7l.manylinux2014_armv7l.whl", hash = "sha256:4f01a5d6444a3258b00dc07b6ea4733e26f8072b788bef750baa37b370266137"}, + {file = "rpds_py-0.26.0-pp39-pypy39_pp73-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl", hash = "sha256:b6e2c12160c72aeda9d1283e612f68804621f448145a210f1bf1d79151c47090"}, + {file = "rpds_py-0.26.0-pp39-pypy39_pp73-manylinux_2_17_s390x.manylinux2014_s390x.whl", hash = "sha256:cb28c1f569f8d33b2b5dcd05d0e6ef7005d8639c54c2f0be824f05aedf715255"}, + {file = "rpds_py-0.26.0-pp39-pypy39_pp73-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:1766b5724c3f779317d5321664a343c07773c8c5fd1532e4039e6cc7d1a815be"}, + {file = "rpds_py-0.26.0-pp39-pypy39_pp73-manylinux_2_5_i686.manylinux1_i686.whl", hash = "sha256:b6d9e5a2ed9c4988c8f9b28b3bc0e3e5b1aaa10c28d210a594ff3a8c02742daf"}, + {file = "rpds_py-0.26.0-pp39-pypy39_pp73-musllinux_1_2_aarch64.whl", hash = "sha256:b5f7a446ddaf6ca0fad9a5535b56fbfc29998bf0e0b450d174bbec0d600e1d72"}, + {file = "rpds_py-0.26.0-pp39-pypy39_pp73-musllinux_1_2_i686.whl", hash = "sha256:eed5ac260dd545fbc20da5f4f15e7efe36a55e0e7cf706e4ec005b491a9546a0"}, + {file = "rpds_py-0.26.0-pp39-pypy39_pp73-musllinux_1_2_x86_64.whl", hash = "sha256:582462833ba7cee52e968b0341b85e392ae53d44c0f9af6a5927c80e539a8b67"}, + {file = "rpds_py-0.26.0-pp39-pypy39_pp73-win_amd64.whl", hash = "sha256:69a607203441e07e9a8a529cff1d5b73f6a160f22db1097211e6212a68567d11"}, + {file = "rpds_py-0.26.0.tar.gz", hash = "sha256:20dae58a859b0906f0685642e591056f1e787f3a8b39c8e8749a45dc7d26bdb0"}, ] [[package]] @@ -4020,27 +4093,27 @@ files = [ [[package]] name = "smart-open" -version = "7.1.0" -description = "Utils for streaming large files (S3, HDFS, GCS, Azure Blob Storage, gzip, bz2...)" +version = "7.3.0.post1" +description = "Utils for streaming large files (S3, HDFS, GCS, SFTP, Azure Blob Storage, gzip, bz2, zst...)" optional = false -python-versions = "<4.0,>=3.7" +python-versions = "<4.0,>=3.8" groups = ["main"] files = [ - {file = "smart_open-7.1.0-py3-none-any.whl", hash = "sha256:4b8489bb6058196258bafe901730c7db0dcf4f083f316e97269c66f45502055b"}, - {file = "smart_open-7.1.0.tar.gz", hash = "sha256:a4f09f84f0f6d3637c6543aca7b5487438877a21360e7368ccf1f704789752ba"}, + {file = "smart_open-7.3.0.post1-py3-none-any.whl", hash = "sha256:c73661a2c24bf045c1e04e08fffc585b59af023fe783d57896f590489db66fb4"}, + {file = "smart_open-7.3.0.post1.tar.gz", hash = "sha256:ce6a3d9bc1afbf6234ad13c010b77f8cd36d24636811e3c52c3b5160f5214d1e"}, ] [package.dependencies] wrapt = "*" [package.extras] -all = ["azure-common", "azure-core", "azure-storage-blob", "boto3", "google-cloud-storage (>=2.6.0)", "paramiko", "requests", "zstandard"] +all = ["smart_open[azure,gcs,http,s3,ssh,webhdfs,zst]"] azure = ["azure-common", "azure-core", "azure-storage-blob"] gcs = ["google-cloud-storage (>=2.6.0)"] http = ["requests"] s3 = ["boto3"] ssh = ["paramiko"] -test = ["awscli", "azure-common", "azure-core", "azure-storage-blob", "boto3", "google-cloud-storage (>=2.6.0)", "moto[server]", "numpy", "paramiko", "pyopenssl", "pytest", "pytest-benchmark", "pytest-rerunfailures", "requests", "responses", "zstandard"] +test = ["awscli", "moto[server]", "numpy", "pyopenssl", "pytest", "pytest-rerunfailures", "pytest_benchmark", "responses", "smart_open[all]"] webhdfs = ["requests"] zst = ["zstandard"] @@ -4351,26 +4424,26 @@ test = ["argcomplete (>=3.0.3)", "mypy (>=1.7.0)", "pre-commit", "pytest (>=7.0, [[package]] name = "types-python-dateutil" -version = "2.9.0.20250516" +version = "2.9.0.20250708" description = "Typing stubs for python-dateutil" optional = false python-versions = ">=3.9" groups = ["test"] files = [ - {file = "types_python_dateutil-2.9.0.20250516-py3-none-any.whl", hash = "sha256:2b2b3f57f9c6a61fba26a9c0ffb9ea5681c9b83e69cd897c6b5f668d9c0cab93"}, - {file = "types_python_dateutil-2.9.0.20250516.tar.gz", hash = "sha256:13e80d6c9c47df23ad773d54b2826bd52dbbb41be87c3f339381c1700ad21ee5"}, + {file = "types_python_dateutil-2.9.0.20250708-py3-none-any.whl", hash = "sha256:4d6d0cc1cc4d24a2dc3816024e502564094497b713f7befda4d5bc7a8e3fd21f"}, + {file = "types_python_dateutil-2.9.0.20250708.tar.gz", hash = "sha256:ccdbd75dab2d6c9696c350579f34cffe2c281e4c5f27a585b2a2438dd1d5c8ab"}, ] [[package]] name = "typing-extensions" -version = "4.14.0" +version = "4.14.1" description = "Backported and Experimental Type Hints for Python 3.9+" optional = false python-versions = ">=3.9" groups = ["main", "test"] files = [ - {file = "typing_extensions-4.14.0-py3-none-any.whl", hash = "sha256:a1514509136dd0b477638fc68d6a91497af5076466ad0fa6c338e44e359944af"}, - {file = "typing_extensions-4.14.0.tar.gz", hash = "sha256:8676b788e32f02ab42d9e7c61324048ae4c6d844a399eebace3d4979d75ceef4"}, + {file = "typing_extensions-4.14.1-py3-none-any.whl", hash = "sha256:d1e1e3b58374dc93031d6eda2420a48ea44a36c2b4766a4fdeb3710755731d76"}, + {file = "typing_extensions-4.14.1.tar.gz", hash = "sha256:38b39f4aeeab64884ce9f74c94263ef78f3c22467c8724005483154c26648d36"}, ] [[package]] @@ -4417,14 +4490,14 @@ dev = ["flake8", "flake8-annotations", "flake8-bandit", "flake8-bugbear", "flake [[package]] name = "urllib3" -version = "2.4.0" +version = "2.5.0" description = "HTTP library with thread-safe connection pooling, file post, and more." optional = false python-versions = ">=3.9" groups = ["main", "docs", "test"] files = [ - {file = "urllib3-2.4.0-py3-none-any.whl", hash = "sha256:4e16665048960a0900c702d4a66415956a584919c03361cac9f1df5c5dd7e813"}, - {file = "urllib3-2.4.0.tar.gz", hash = "sha256:414bc6535b787febd7567804cc015fee39daab8ad86268f1310a9250697de466"}, + {file = "urllib3-2.5.0-py3-none-any.whl", hash = "sha256:e6b01673c0fa6a13e374b50871808eb3bf7046c4b125b216f6bf1cc604cff0dc"}, + {file = "urllib3-2.5.0.tar.gz", hash = "sha256:3fc47733c7e419d4bc3f6b3dc2b4f890bb743906a30d56ba4a5bfa4bbff92760"}, ] [package.extras] @@ -4435,14 +4508,14 @@ zstd = ["zstandard (>=0.18.0)"] [[package]] name = "virtualenv" -version = "20.31.2" +version = "20.35.1" description = "Virtual Python Environment builder" optional = false python-versions = ">=3.8" groups = ["main"] files = [ - {file = "virtualenv-20.31.2-py3-none-any.whl", hash = "sha256:36efd0d9650ee985f0cad72065001e66d49a6f24eb44d98980f630686243cf11"}, - {file = "virtualenv-20.31.2.tar.gz", hash = "sha256:e10c0a9d02835e592521be48b332b6caee6887f332c111aa79a09b9e79efc2af"}, + {file = "virtualenv-20.35.1-py3-none-any.whl", hash = "sha256:1d9d93cd01d35b785476e2fa7af711a98d40d227a078941695bbae394f8737e2"}, + {file = "virtualenv-20.35.1.tar.gz", hash = "sha256:041dac43b6899858a91838b616599e80000e545dee01a21172a6a46746472cb2"}, ] [package.dependencies] @@ -4750,4 +4823,4 @@ type = ["pytest-mypy"] [metadata] lock-version = "2.1" python-versions = "^3.11" -content-hash = "6720576cf9ff57c7bb15b97e268bb414218f6a053e7e0a5bdd45d022c0847111" +content-hash = "9a052e3a816450844fa2cf3427e4660715977aca3b14561d1e9991899624b7c2" diff --git a/pyproject.toml b/pyproject.toml index 71610478..6d13354e 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -33,6 +33,14 @@ cryptography = "43.0.3" executing = "1.2.0" pydantic = ">= 2.10.6" ipywidgets = "8.1.2" +python-client = { git = "https://github.com/ray-project/kuberay.git", subdirectory = "clients/python-client", rev = "b2fd91b58c2bbe22f9b4f730c5a8f3180c05e570" } + +[[tool.poetry.source]] +name = "pypi" + +[[tool.poetry.source]] +name = "testpypi" +url = "https://test.pypi.org/simple/" [tool.poetry.group.docs] optional = true @@ -51,6 +59,10 @@ pytest-mock = "3.11.1" pytest-timeout = "2.3.1" jupyterlab = "4.3.1" + +[tool.poetry.group.dev.dependencies] +diff-cover = "^9.6.0" + [tool.pytest.ini_options] filterwarnings = [ "ignore::DeprecationWarning:pkg_resources", diff --git a/src/codeflare_sdk/__init__.py b/src/codeflare_sdk/__init__.py index 9ab5c745..a27702e7 100644 --- a/src/codeflare_sdk/__init__.py +++ b/src/codeflare_sdk/__init__.py @@ -10,6 +10,8 @@ AWManager, AppWrapperStatus, RayJobClient, + RayJob, + ManagedClusterConfig, ) from .common.widgets import view_clusters diff --git a/src/codeflare_sdk/common/kueue/kueue.py b/src/codeflare_sdk/common/kueue/kueue.py index 00f3364a..a721713e 100644 --- a/src/codeflare_sdk/common/kueue/kueue.py +++ b/src/codeflare_sdk/common/kueue/kueue.py @@ -18,6 +18,8 @@ from kubernetes import client from kubernetes.client.exceptions import ApiException +from ...common.utils import get_current_namespace + def get_default_kueue_name(namespace: str) -> Optional[str]: """ @@ -81,7 +83,6 @@ def list_local_queues( List[dict]: A list of dictionaries containing the name of the local queue and the available flavors """ - from ...ray.cluster.cluster import get_current_namespace if namespace is None: # pragma: no cover namespace = get_current_namespace() diff --git a/src/codeflare_sdk/common/utils/__init__.py b/src/codeflare_sdk/common/utils/__init__.py index e69de29b..e662bf5e 100644 --- a/src/codeflare_sdk/common/utils/__init__.py +++ b/src/codeflare_sdk/common/utils/__init__.py @@ -0,0 +1,7 @@ +""" +Common utilities for the CodeFlare SDK. +""" + +from .k8s_utils import get_current_namespace + +__all__ = ["get_current_namespace"] diff --git a/src/codeflare_sdk/common/utils/constants.py b/src/codeflare_sdk/common/utils/constants.py index fcd064d6..00559d2e 100644 --- a/src/codeflare_sdk/common/utils/constants.py +++ b/src/codeflare_sdk/common/utils/constants.py @@ -12,3 +12,4 @@ "3.11": CUDA_PY311_RUNTIME_IMAGE, "3.12": CUDA_PY312_RUNTIME_IMAGE, } +MOUNT_PATH = "/home/ray/files" diff --git a/src/codeflare_sdk/common/utils/k8s_utils.py b/src/codeflare_sdk/common/utils/k8s_utils.py new file mode 100644 index 00000000..e2e03a5d --- /dev/null +++ b/src/codeflare_sdk/common/utils/k8s_utils.py @@ -0,0 +1,33 @@ +""" +Kubernetes utility functions for the CodeFlare SDK. +""" + +import os +from kubernetes import config +from ..kubernetes_cluster import config_check, _kube_api_error_handling + + +def get_current_namespace(): # pragma: no cover + """ + Retrieves the current Kubernetes namespace. + + Returns: + str: + The current namespace or None if not found. + """ + if os.path.isfile("/var/run/secrets/kubernetes.io/serviceaccount/namespace"): + try: + file = open("/var/run/secrets/kubernetes.io/serviceaccount/namespace", "r") + active_context = file.readline().strip("\n") + return active_context + except Exception as e: + print("Unable to find current namespace") + print("trying to gather from current context") + try: + _, active_context = config.list_kube_config_contexts(config_check()) + except Exception as e: + return _kube_api_error_handling(e) + try: + return active_context["context"]["namespace"] + except KeyError: + return None diff --git a/src/codeflare_sdk/common/utils/test_demos.py b/src/codeflare_sdk/common/utils/test_demos.py new file mode 100644 index 00000000..9124cbec --- /dev/null +++ b/src/codeflare_sdk/common/utils/test_demos.py @@ -0,0 +1,57 @@ +# Copyright 2025 IBM, Red Hat +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +""" +Tests for demos module. +""" + +import pytest +import tempfile +from pathlib import Path +from unittest.mock import patch, MagicMock +from codeflare_sdk.common.utils.demos import copy_demo_nbs + + +class TestCopyDemoNbs: + """Test cases for copy_demo_nbs function.""" + + def test_copy_demo_nbs_directory_exists_error(self): + """Test that FileExistsError is raised when directory exists and overwrite=False.""" + with tempfile.TemporaryDirectory() as temp_dir: + # Create a subdirectory that will conflict + conflict_dir = Path(temp_dir) / "demo-notebooks" + conflict_dir.mkdir() + + with pytest.raises(FileExistsError, match="Directory.*already exists"): + copy_demo_nbs(dir=str(conflict_dir), overwrite=False) + + def test_copy_demo_nbs_overwrite_true(self): + """Test that overwrite=True allows copying to existing directory.""" + with tempfile.TemporaryDirectory() as temp_dir: + # Create a subdirectory that will conflict + conflict_dir = Path(temp_dir) / "demo-notebooks" + conflict_dir.mkdir() + + # Mock the demo_dir to point to a real directory + with patch("codeflare_sdk.common.utils.demos.demo_dir", temp_dir): + # Should not raise an error with overwrite=True + copy_demo_nbs(dir=str(conflict_dir), overwrite=True) + + def test_copy_demo_nbs_default_parameters(self): + """Test copy_demo_nbs with default parameters.""" + with tempfile.TemporaryDirectory() as temp_dir: + # Mock the demo_dir to point to a real directory + with patch("codeflare_sdk.common.utils.demos.demo_dir", temp_dir): + # Should work with default parameters + copy_demo_nbs(dir=temp_dir, overwrite=True) diff --git a/src/codeflare_sdk/common/utils/test_k8s_utils.py b/src/codeflare_sdk/common/utils/test_k8s_utils.py new file mode 100644 index 00000000..fcd0623d --- /dev/null +++ b/src/codeflare_sdk/common/utils/test_k8s_utils.py @@ -0,0 +1,255 @@ +# Copyright 2025 IBM, Red Hat +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +""" +Tests for k8s_utils module. +""" + +import pytest +from unittest.mock import mock_open, patch, MagicMock +from codeflare_sdk.common.utils.k8s_utils import get_current_namespace + + +class TestGetCurrentNamespace: + """Test cases for get_current_namespace function.""" + + def test_get_current_namespace_incluster_success(self): + """Test successful namespace detection from in-cluster service account.""" + mock_file_content = "test-namespace\n" + + with patch("os.path.isfile", return_value=True): + with patch("builtins.open", mock_open(read_data=mock_file_content)): + result = get_current_namespace() + + assert result == "test-namespace" + + def test_get_current_namespace_incluster_file_read_error(self): + """Test handling of file read errors when reading service account namespace.""" + with patch("os.path.isfile", return_value=True): + with patch("builtins.open", side_effect=IOError("File read error")): + with patch("builtins.print") as mock_print: + # Mock config_check to avoid kubeconfig fallback + with patch( + "codeflare_sdk.common.utils.k8s_utils.config_check", + side_effect=Exception("Config error"), + ): + with patch( + "codeflare_sdk.common.utils.k8s_utils._kube_api_error_handling", + return_value=None, + ): + result = get_current_namespace() + + assert result is None + # Should see both error messages: in-cluster failure and kubeconfig fallback + mock_print.assert_any_call("Unable to find current namespace") + mock_print.assert_any_call("trying to gather from current context") + + def test_get_current_namespace_incluster_file_open_error(self): + """Test handling of file open errors when reading service account namespace.""" + with patch("os.path.isfile", return_value=True): + with patch( + "builtins.open", side_effect=PermissionError("Permission denied") + ): + with patch("builtins.print") as mock_print: + # Mock config_check to avoid kubeconfig fallback + with patch( + "codeflare_sdk.common.utils.k8s_utils.config_check", + side_effect=Exception("Config error"), + ): + with patch( + "codeflare_sdk.common.utils.k8s_utils._kube_api_error_handling", + return_value=None, + ): + result = get_current_namespace() + + assert result is None + # Should see both error messages: in-cluster failure and kubeconfig fallback + mock_print.assert_any_call("Unable to find current namespace") + mock_print.assert_any_call("trying to gather from current context") + + def test_get_current_namespace_kubeconfig_success(self): + """Test successful namespace detection from kubeconfig context.""" + mock_contexts = [ + {"name": "context1", "context": {"namespace": "default"}}, + {"name": "context2", "context": {"namespace": "test-namespace"}}, + ] + mock_active_context = { + "name": "context2", + "context": {"namespace": "test-namespace"}, + } + + with patch("os.path.isfile", return_value=False): + with patch("builtins.print") as mock_print: + with patch( + "codeflare_sdk.common.utils.k8s_utils.config_check", + return_value="~/.kube/config", + ): + with patch( + "kubernetes.config.list_kube_config_contexts", + return_value=(mock_contexts, mock_active_context), + ): + result = get_current_namespace() + + assert result == "test-namespace" + mock_print.assert_called_with("trying to gather from current context") + + def test_get_current_namespace_kubeconfig_no_namespace_in_context(self): + """Test handling when kubeconfig context has no namespace field.""" + mock_contexts = [ + {"name": "context1", "context": {}}, + {"name": "context2", "context": {"cluster": "test-cluster"}}, + ] + mock_active_context = { + "name": "context2", + "context": {"cluster": "test-cluster"}, + } + + with patch("os.path.isfile", return_value=False): + with patch("builtins.print") as mock_print: + with patch( + "codeflare_sdk.common.utils.k8s_utils.config_check", + return_value="~/.kube/config", + ): + with patch( + "kubernetes.config.list_kube_config_contexts", + return_value=(mock_contexts, mock_active_context), + ): + result = get_current_namespace() + + assert result is None + mock_print.assert_called_with("trying to gather from current context") + + def test_get_current_namespace_kubeconfig_config_check_error(self): + """Test handling when config_check raises an exception.""" + with patch("os.path.isfile", return_value=False): + with patch("builtins.print") as mock_print: + with patch( + "codeflare_sdk.common.utils.k8s_utils.config_check", + side_effect=Exception("Config error"), + ): + with patch( + "codeflare_sdk.common.utils.k8s_utils._kube_api_error_handling", + return_value=None, + ) as mock_error_handler: + result = get_current_namespace() + + assert result is None + mock_print.assert_called_with("trying to gather from current context") + mock_error_handler.assert_called_once() + + def test_get_current_namespace_kubeconfig_list_contexts_error(self): + """Test handling when list_kube_config_contexts raises an exception.""" + with patch("os.path.isfile", return_value=False): + with patch("builtins.print") as mock_print: + with patch( + "codeflare_sdk.common.utils.k8s_utils.config_check", + return_value="~/.kube/config", + ): + with patch( + "kubernetes.config.list_kube_config_contexts", + side_effect=Exception("Context error"), + ): + with patch( + "codeflare_sdk.common.utils.k8s_utils._kube_api_error_handling", + return_value=None, + ) as mock_error_handler: + result = get_current_namespace() + + assert result is None + mock_print.assert_called_with("trying to gather from current context") + mock_error_handler.assert_called_once() + + def test_get_current_namespace_kubeconfig_key_error(self): + """Test handling when accessing context namespace raises KeyError.""" + mock_contexts = [{"name": "context1", "context": {"namespace": "default"}}] + mock_active_context = {"name": "context1"} # Missing 'context' key + + with patch("os.path.isfile", return_value=False): + with patch("builtins.print") as mock_print: + with patch( + "codeflare_sdk.common.utils.k8s_utils.config_check", + return_value="~/.kube/config", + ): + with patch( + "kubernetes.config.list_kube_config_contexts", + return_value=(mock_contexts, mock_active_context), + ): + result = get_current_namespace() + + assert result is None + mock_print.assert_called_with("trying to gather from current context") + + def test_get_current_namespace_fallback_flow(self): + """Test the complete fallback flow from in-cluster to kubeconfig.""" + # First attempt: in-cluster file doesn't exist + # Second attempt: kubeconfig context has namespace + mock_contexts = [ + {"name": "context1", "context": {"namespace": "fallback-namespace"}} + ] + mock_active_context = { + "name": "context1", + "context": {"namespace": "fallback-namespace"}, + } + + with patch("os.path.isfile", return_value=False): + with patch("builtins.print") as mock_print: + with patch( + "codeflare_sdk.common.utils.k8s_utils.config_check", + return_value="~/.kube/config", + ): + with patch( + "kubernetes.config.list_kube_config_contexts", + return_value=(mock_contexts, mock_active_context), + ): + result = get_current_namespace() + + assert result == "fallback-namespace" + mock_print.assert_called_with("trying to gather from current context") + + def test_get_current_namespace_complete_failure(self): + """Test complete failure scenario where no namespace can be detected.""" + with patch("os.path.isfile", return_value=False): + with patch("builtins.print") as mock_print: + with patch( + "codeflare_sdk.common.utils.k8s_utils.config_check", + side_effect=Exception("Config error"), + ): + with patch( + "codeflare_sdk.common.utils.k8s_utils._kube_api_error_handling", + return_value=None, + ): + result = get_current_namespace() + + assert result is None + mock_print.assert_called_with("trying to gather from current context") + + def test_get_current_namespace_mixed_errors(self): + """Test scenario with mixed error conditions.""" + # In-cluster file exists but read fails, then kubeconfig also fails + with patch("os.path.isfile", return_value=True): + with patch("builtins.open", side_effect=IOError("File read error")): + with patch("builtins.print") as mock_print: + with patch( + "codeflare_sdk.common.utils.k8s_utils.config_check", + side_effect=Exception("Config error"), + ): + with patch( + "codeflare_sdk.common.utils.k8s_utils._kube_api_error_handling", + return_value=None, + ): + result = get_current_namespace() + + assert result is None + # Should see both error messages + assert mock_print.call_count >= 2 diff --git a/src/codeflare_sdk/common/utils/test_utils.py b/src/codeflare_sdk/common/utils/test_utils.py new file mode 100644 index 00000000..d330bc4d --- /dev/null +++ b/src/codeflare_sdk/common/utils/test_utils.py @@ -0,0 +1,209 @@ +# Copyright 2025 IBM, Red Hat +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +""" +Tests for common/utils/utils.py +""" + +import pytest +from collections import namedtuple +from codeflare_sdk.common.utils.utils import ( + update_image, + get_ray_image_for_python_version, +) +from codeflare_sdk.common.utils.constants import ( + SUPPORTED_PYTHON_VERSIONS, + CUDA_PY311_RUNTIME_IMAGE, + CUDA_PY312_RUNTIME_IMAGE, +) + + +def test_update_image_with_empty_string_python_311(mocker): + """Test that update_image() with empty string returns default image for Python 3.11.""" + # Mock sys.version_info to simulate Python 3.11 + VersionInfo = namedtuple( + "version_info", ["major", "minor", "micro", "releaselevel", "serial"] + ) + mocker.patch("sys.version_info", VersionInfo(3, 11, 0, "final", 0)) + + # Test with empty image (should use default for Python 3.11) + image = update_image("") + assert image == CUDA_PY311_RUNTIME_IMAGE + assert image == SUPPORTED_PYTHON_VERSIONS["3.11"] + + +def test_update_image_with_empty_string_python_312(mocker): + """Test that update_image() with empty string returns default image for Python 3.12.""" + # Mock sys.version_info to simulate Python 3.12 + VersionInfo = namedtuple( + "version_info", ["major", "minor", "micro", "releaselevel", "serial"] + ) + mocker.patch("sys.version_info", VersionInfo(3, 12, 0, "final", 0)) + + # Test with empty image (should use default for Python 3.12) + image = update_image("") + assert image == CUDA_PY312_RUNTIME_IMAGE + assert image == SUPPORTED_PYTHON_VERSIONS["3.12"] + + +def test_update_image_with_none_python_311(mocker): + """Test that update_image() with None returns default image for Python 3.11.""" + # Mock sys.version_info to simulate Python 3.11 + VersionInfo = namedtuple( + "version_info", ["major", "minor", "micro", "releaselevel", "serial"] + ) + mocker.patch("sys.version_info", VersionInfo(3, 11, 0, "final", 0)) + + # Test with None image (should use default for Python 3.11) + image = update_image(None) + assert image == CUDA_PY311_RUNTIME_IMAGE + + +def test_update_image_with_none_python_312(mocker): + """Test that update_image() with None returns default image for Python 3.12.""" + # Mock sys.version_info to simulate Python 3.12 + VersionInfo = namedtuple( + "version_info", ["major", "minor", "micro", "releaselevel", "serial"] + ) + mocker.patch("sys.version_info", VersionInfo(3, 12, 0, "final", 0)) + + # Test with None image (should use default for Python 3.12) + image = update_image(None) + assert image == CUDA_PY312_RUNTIME_IMAGE + + +def test_update_image_with_unsupported_python_version(mocker): + """Test update_image() warning for unsupported Python versions.""" + # Mock sys.version_info to simulate Python 3.8 (unsupported) + VersionInfo = namedtuple( + "version_info", ["major", "minor", "micro", "releaselevel", "serial"] + ) + mocker.patch("sys.version_info", VersionInfo(3, 8, 0, "final", 0)) + + # Mock warnings.warn to check if it gets called + warn_mock = mocker.patch("warnings.warn") + + # Call update_image with empty image + image = update_image("") + + # Assert that the warning was called with the expected message + warn_mock.assert_called_once() + assert "No default Ray image defined for 3.8" in warn_mock.call_args[0][0] + assert "3.11, 3.12" in warn_mock.call_args[0][0] + + # Assert that no image was set since the Python version is not supported + assert image is None + + +def test_update_image_with_provided_custom_image(): + """Test that providing a custom image bypasses auto-detection.""" + custom_image = "my-custom-ray:latest" + image = update_image(custom_image) + + # Should return the provided image unchanged + assert image == custom_image + + +def test_update_image_with_provided_image_empty_string(): + """Test update_image() with provided custom image as a non-empty string.""" + custom_image = "docker.io/rayproject/ray:2.40.0" + image = update_image(custom_image) + + # Should return the provided image unchanged + assert image == custom_image + + +def test_get_ray_image_for_python_version_explicit_311(): + """Test get_ray_image_for_python_version() with explicit Python 3.11.""" + image = get_ray_image_for_python_version("3.11") + assert image == CUDA_PY311_RUNTIME_IMAGE + + +def test_get_ray_image_for_python_version_explicit_312(): + """Test get_ray_image_for_python_version() with explicit Python 3.12.""" + image = get_ray_image_for_python_version("3.12") + assert image == CUDA_PY312_RUNTIME_IMAGE + + +def test_get_ray_image_for_python_version_auto_detect_311(mocker): + """Test get_ray_image_for_python_version() auto-detects Python 3.11.""" + # Mock sys.version_info to simulate Python 3.11 + VersionInfo = namedtuple( + "version_info", ["major", "minor", "micro", "releaselevel", "serial"] + ) + mocker.patch("sys.version_info", VersionInfo(3, 11, 0, "final", 0)) + + # Test with None (should auto-detect) + image = get_ray_image_for_python_version() + assert image == CUDA_PY311_RUNTIME_IMAGE + + +def test_get_ray_image_for_python_version_auto_detect_312(mocker): + """Test get_ray_image_for_python_version() auto-detects Python 3.12.""" + # Mock sys.version_info to simulate Python 3.12 + VersionInfo = namedtuple( + "version_info", ["major", "minor", "micro", "releaselevel", "serial"] + ) + mocker.patch("sys.version_info", VersionInfo(3, 12, 0, "final", 0)) + + # Test with None (should auto-detect) + image = get_ray_image_for_python_version() + assert image == CUDA_PY312_RUNTIME_IMAGE + + +def test_get_ray_image_for_python_version_unsupported_with_warning(mocker): + """Test get_ray_image_for_python_version() warns for unsupported versions.""" + warn_mock = mocker.patch("warnings.warn") + + # Test with unsupported version and warn_on_unsupported=True (default) + image = get_ray_image_for_python_version("3.9", warn_on_unsupported=True) + + # Should have warned + warn_mock.assert_called_once() + assert "No default Ray image defined for 3.9" in warn_mock.call_args[0][0] + + # Should return None + assert image is None + + +def test_get_ray_image_for_python_version_unsupported_without_warning(): + """Test get_ray_image_for_python_version() falls back to 3.12 without warning.""" + # Test with unsupported version and warn_on_unsupported=False + image = get_ray_image_for_python_version("3.10", warn_on_unsupported=False) + + # Should fall back to Python 3.12 image + assert image == CUDA_PY312_RUNTIME_IMAGE + + +def test_get_ray_image_for_python_version_unsupported_silent_fallback(): + """Test get_ray_image_for_python_version() silently falls back for old versions.""" + # Test with Python 3.8 and warn_on_unsupported=False + image = get_ray_image_for_python_version("3.8", warn_on_unsupported=False) + + # Should fall back to Python 3.12 image without warning + assert image == CUDA_PY312_RUNTIME_IMAGE + + +def test_get_ray_image_for_python_version_none_defaults_to_current(mocker): + """Test that passing None to get_ray_image_for_python_version() uses current Python.""" + # Mock sys.version_info to simulate Python 3.11 + VersionInfo = namedtuple( + "version_info", ["major", "minor", "micro", "releaselevel", "serial"] + ) + mocker.patch("sys.version_info", VersionInfo(3, 11, 5, "final", 0)) + + # Passing None should detect the mocked version + image = get_ray_image_for_python_version(None, warn_on_unsupported=True) + + assert image == CUDA_PY311_RUNTIME_IMAGE diff --git a/src/codeflare_sdk/common/utils/test_validation.py b/src/codeflare_sdk/common/utils/test_validation.py new file mode 100644 index 00000000..20416d00 --- /dev/null +++ b/src/codeflare_sdk/common/utils/test_validation.py @@ -0,0 +1,224 @@ +# Copyright 2022-2025 IBM, Red Hat +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +from codeflare_sdk.common.utils.validation import ( + extract_ray_version_from_image, + validate_ray_version_compatibility, +) +from codeflare_sdk.common.utils.constants import RAY_VERSION + + +class TestRayVersionDetection: + """Test Ray version detection from container image names.""" + + def test_extract_ray_version_standard_format(self): + """Test extraction from standard Ray image formats.""" + # Standard format + assert extract_ray_version_from_image("ray:2.47.1") == "2.47.1" + assert extract_ray_version_from_image("ray:2.46.0") == "2.46.0" + assert extract_ray_version_from_image("ray:1.13.0") == "1.13.0" + + def test_extract_ray_version_with_registry(self): + """Test extraction from images with registry prefixes.""" + assert extract_ray_version_from_image("quay.io/ray:2.47.1") == "2.47.1" + assert ( + extract_ray_version_from_image("docker.io/rayproject/ray:2.47.1") + == "2.47.1" + ) + assert ( + extract_ray_version_from_image("gcr.io/my-project/ray:2.47.1") == "2.47.1" + ) + + def test_extract_ray_version_with_suffixes(self): + """Test extraction from images with version suffixes.""" + assert ( + extract_ray_version_from_image("quay.io/modh/ray:2.47.1-py311-cu121") + == "2.47.1" + ) + assert extract_ray_version_from_image("ray:2.47.1-py311") == "2.47.1" + assert extract_ray_version_from_image("ray:2.47.1-gpu") == "2.47.1" + assert extract_ray_version_from_image("ray:2.47.1-rocm62") == "2.47.1" + + def test_extract_ray_version_complex_registry_paths(self): + """Test extraction from complex registry paths.""" + assert ( + extract_ray_version_from_image("quay.io/modh/ray:2.47.1-py311-cu121") + == "2.47.1" + ) + assert ( + extract_ray_version_from_image("registry.company.com/team/ray:2.47.1") + == "2.47.1" + ) + + def test_extract_ray_version_no_version_found(self): + """Test cases where no version can be extracted.""" + # SHA-based tags + assert ( + extract_ray_version_from_image( + "quay.io/modh/ray@sha256:6d076aeb38ab3c34a6a2ef0f58dc667089aa15826fa08a73273c629333e12f1e" + ) + is None + ) + + # Non-semantic versions + assert extract_ray_version_from_image("ray:latest") is None + assert extract_ray_version_from_image("ray:nightly") is None + assert ( + extract_ray_version_from_image("ray:v2.47") is None + ) # Missing patch version + + # Non-Ray images + assert extract_ray_version_from_image("python:3.11") is None + assert extract_ray_version_from_image("ubuntu:20.04") is None + + # Empty or None + assert extract_ray_version_from_image("") is None + assert extract_ray_version_from_image(None) is None + + def test_extract_ray_version_edge_cases(self): + """Test edge cases for version extraction.""" + # Version with 'v' prefix should not match our pattern + assert extract_ray_version_from_image("ray:v2.47.1") is None + + # Multiple version-like patterns - should match the first valid one + assert ( + extract_ray_version_from_image("registry/ray:2.47.1-based-on-1.0.0") + == "2.47.1" + ) + + +class TestRayVersionValidation: + """Test Ray version compatibility validation.""" + + def test_validate_compatible_versions(self): + """Test validation with compatible Ray versions.""" + # Exact match + is_compatible, is_warning, message = validate_ray_version_compatibility( + f"ray:{RAY_VERSION}" + ) + assert is_compatible is True + assert is_warning is False + assert "Ray versions match" in message + + # With registry and suffixes + is_compatible, is_warning, message = validate_ray_version_compatibility( + f"quay.io/modh/ray:{RAY_VERSION}-py311-cu121" + ) + assert is_compatible is True + assert is_warning is False + assert "Ray versions match" in message + + def test_validate_incompatible_versions(self): + """Test validation with incompatible Ray versions.""" + # Different version + is_compatible, is_warning, message = validate_ray_version_compatibility( + "ray:2.46.0" + ) + assert is_compatible is False + assert is_warning is False + assert "Ray version mismatch detected" in message + assert "CodeFlare SDK uses Ray" in message + assert "runtime image uses Ray" in message + + # Older version + is_compatible, is_warning, message = validate_ray_version_compatibility( + "ray:1.13.0" + ) + assert is_compatible is False + assert is_warning is False + assert "Ray version mismatch detected" in message + + def test_validate_empty_image(self): + """Test validation with no custom image (should use default).""" + # Empty string + is_compatible, is_warning, message = validate_ray_version_compatibility("") + assert is_compatible is True + assert is_warning is False + assert "Using default Ray image compatible with SDK" in message + + # None + is_compatible, is_warning, message = validate_ray_version_compatibility(None) + assert is_compatible is True + assert is_warning is False + assert "Using default Ray image compatible with SDK" in message + + def test_validate_unknown_version(self): + """Test validation when version cannot be determined.""" + # SHA-based image + is_compatible, is_warning, message = validate_ray_version_compatibility( + "quay.io/modh/ray@sha256:6d076aeb38ab3c34a6a2ef0f58dc667089aa15826fa08a73273c629333e12f1e" + ) + assert is_compatible is True + assert is_warning is True + assert "Cannot determine Ray version" in message + + # Custom image without version + is_compatible, is_warning, message = validate_ray_version_compatibility( + "my-custom-ray:latest" + ) + assert is_compatible is True + assert is_warning is True + assert "Cannot determine Ray version" in message + + def test_validate_custom_sdk_version(self): + """Test validation with custom SDK version.""" + # Compatible with custom SDK version + is_compatible, is_warning, message = validate_ray_version_compatibility( + "ray:2.46.0", "2.46.0" + ) + assert is_compatible is True + assert is_warning is False + assert "Ray versions match" in message + + # Incompatible with custom SDK version + is_compatible, is_warning, message = validate_ray_version_compatibility( + "ray:2.47.1", "2.46.0" + ) + assert is_compatible is False + assert is_warning is False + assert "CodeFlare SDK uses Ray 2.46.0" in message + assert "runtime image uses Ray 2.47.1" in message + + def test_validate_message_content(self): + """Test that validation messages contain expected guidance.""" + # Mismatch message should contain helpful guidance + is_compatible, is_warning, message = validate_ray_version_compatibility( + "ray:2.46.0" + ) + assert is_compatible is False + assert is_warning is False + assert "compatibility issues" in message.lower() + assert "unexpected behavior" in message.lower() + assert "please use a runtime image" in message.lower() + assert "update your sdk version" in message.lower() + + def test_semantic_version_comparison(self): + """Test that semantic version comparison works correctly.""" + # Test that 2.10.0 > 2.9.1 (would fail with string comparison) + is_compatible, is_warning, message = validate_ray_version_compatibility( + "ray:2.10.0", "2.9.1" + ) + assert is_compatible is False + assert is_warning is False + assert "CodeFlare SDK uses Ray 2.9.1" in message + assert "runtime image uses Ray 2.10.0" in message + + # Test that 2.9.1 < 2.10.0 (would fail with string comparison) + is_compatible, is_warning, message = validate_ray_version_compatibility( + "ray:2.9.1", "2.10.0" + ) + assert is_compatible is False + assert is_warning is False + assert "CodeFlare SDK uses Ray 2.10.0" in message + assert "runtime image uses Ray 2.9.1" in message diff --git a/src/codeflare_sdk/common/utils/utils.py b/src/codeflare_sdk/common/utils/utils.py index f876e924..7e30b994 100644 --- a/src/codeflare_sdk/common/utils/utils.py +++ b/src/codeflare_sdk/common/utils/utils.py @@ -19,6 +19,17 @@ ) +def update_image(image) -> str: + """ + The update_image() function automatically sets the image config parameter to a preset image based on Python version if not specified. + This now points to the centralized function in utils.py. + """ + if not image: + # Pull the image based on the matching Python version (or output a warning if not supported) + image = get_ray_image_for_python_version(warn_on_unsupported=True) + return image + + def get_ray_image_for_python_version(python_version=None, warn_on_unsupported=True): """ Get the appropriate Ray image for a given Python version. diff --git a/src/codeflare_sdk/common/utils/validation.py b/src/codeflare_sdk/common/utils/validation.py new file mode 100644 index 00000000..ec749f7c --- /dev/null +++ b/src/codeflare_sdk/common/utils/validation.py @@ -0,0 +1,134 @@ +# Copyright 2022-2025 IBM, Red Hat +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +""" +Validation utilities for the CodeFlare SDK. + +This module contains validation functions used across the SDK for ensuring +configuration compatibility and correctness. +""" + +import logging +import re +from typing import Optional, Tuple +from packaging.version import Version, InvalidVersion +from .constants import RAY_VERSION + +logger = logging.getLogger(__name__) + + +def extract_ray_version_from_image(image_name: str) -> Optional[str]: + """ + Extract Ray version from a container image name. + + Supports various image naming patterns: + - quay.io/modh/ray:2.47.1-py311-cu121 + - ray:2.47.1 + - some-registry/ray:2.47.1-py311 + - quay.io/modh/ray@sha256:... (falls back to None) + + Args: + image_name: The container image name/tag + + Returns: + The extracted Ray version, or None if not found + """ + if not image_name: + return None + + # Pattern to match semantic version after ray: or ray/ + # Looks for patterns like ray:2.47.1, ray:2.47.1-py311, etc. + patterns = [ + r"ray:(\d+\.\d+\.\d+)", # ray:2.47.1 + r"ray/[^:]*:(\d+\.\d+\.\d+)", # registry/ray:2.47.1 + r"/ray:(\d+\.\d+\.\d+)", # any-registry/ray:2.47.1 + ] + + for pattern in patterns: + match = re.search(pattern, image_name) + if match: + return match.group(1) + + # If we can't extract version, return None to indicate unknown + return None + + +def validate_ray_version_compatibility( + image_name: str, sdk_ray_version: str = RAY_VERSION +) -> Tuple[bool, bool, str]: + """ + Validate that the Ray version in the runtime image matches the SDK's Ray version. + + Args: + image_name: The container image name/tag + sdk_ray_version: The Ray version used by the CodeFlare SDK + + Returns: + tuple: (is_compatible, is_warning, message) + - is_compatible: True if versions match or cannot be determined, False if mismatch + - is_warning: True if this is a warning (non-fatal), False otherwise + - message: Descriptive message about the validation result + """ + if not image_name: + # No custom image specified, will use default - this is compatible + logger.debug("Using default Ray image compatible with SDK") + return True, False, "Using default Ray image compatible with SDK" + + image_ray_version = extract_ray_version_from_image(image_name) + + if image_ray_version is None: + # Cannot determine version from image name, issue a warning but allow + return ( + True, + True, + f"Cannot determine Ray version from image '{image_name}'. Please ensure it's compatible with Ray {sdk_ray_version}", + ) + + # Use semantic version comparison for robust version checking + try: + sdk_version = Version(sdk_ray_version) + image_version = Version(image_ray_version) + + if image_version != sdk_version: + # Version mismatch detected + message = ( + f"Ray version mismatch detected!\n" + f"CodeFlare SDK uses Ray {sdk_ray_version}, but runtime image uses Ray {image_ray_version}.\n" + f"This mismatch can cause compatibility issues and unexpected behavior.\n" + f"Please use a runtime image with Ray {sdk_ray_version} or update your SDK version." + ) + return False, False, message + except InvalidVersion as e: + # If version parsing fails, fall back to string comparison with a warning + logger.warning( + f"Failed to parse version for comparison ({e}), falling back to string comparison" + ) + if image_ray_version != sdk_ray_version: + message = ( + f"Ray version mismatch detected!\n" + f"CodeFlare SDK uses Ray {sdk_ray_version}, but runtime image uses Ray {image_ray_version}.\n" + f"This mismatch can cause compatibility issues and unexpected behavior.\n" + f"Please use a runtime image with Ray {sdk_ray_version} or update your SDK version." + ) + return False, False, message + + # Versions match + logger.debug( + f"Ray version validation successful: SDK and runtime image both use Ray {sdk_ray_version}" + ) + return ( + True, + False, + f"Ray versions match: SDK and runtime image both use Ray {sdk_ray_version}", + ) diff --git a/src/codeflare_sdk/common/widgets/test_widgets.py b/src/codeflare_sdk/common/widgets/test_widgets.py index f88d8eb2..55be2b75 100644 --- a/src/codeflare_sdk/common/widgets/test_widgets.py +++ b/src/codeflare_sdk/common/widgets/test_widgets.py @@ -106,7 +106,7 @@ def test_view_clusters(mocker, capsys): # Prepare to run view_clusters when notebook environment is detected mocker.patch("codeflare_sdk.common.widgets.widgets.is_notebook", return_value=True) mock_get_current_namespace = mocker.patch( - "codeflare_sdk.ray.cluster.cluster.get_current_namespace", + "codeflare_sdk.common.widgets.widgets.get_current_namespace", return_value="default", ) namespace = mock_get_current_namespace.return_value @@ -250,7 +250,7 @@ def test_ray_cluster_manager_widgets_init(mocker, capsys): return_value=test_ray_clusters_df, ) mocker.patch( - "codeflare_sdk.ray.cluster.cluster.get_current_namespace", + "codeflare_sdk.common.utils.get_current_namespace", return_value=namespace, ) mock_delete_cluster = mocker.patch( diff --git a/src/codeflare_sdk/common/widgets/widgets.py b/src/codeflare_sdk/common/widgets/widgets.py index 36d896e8..c813fabc 100644 --- a/src/codeflare_sdk/common/widgets/widgets.py +++ b/src/codeflare_sdk/common/widgets/widgets.py @@ -26,6 +26,8 @@ import ipywidgets as widgets from IPython.display import display, HTML, Javascript import pandas as pd + +from ...common.utils import get_current_namespace from ...ray.cluster.config import ClusterConfiguration from ...ray.cluster.status import RayClusterStatus from ..kubernetes_cluster import _kube_api_error_handling @@ -43,8 +45,6 @@ class RayClusterManagerWidgets: """ def __init__(self, ray_clusters_df: pd.DataFrame, namespace: str = None): - from ...ray.cluster.cluster import get_current_namespace - # Data self.ray_clusters_df = ray_clusters_df self.namespace = get_current_namespace() if not namespace else namespace @@ -353,8 +353,6 @@ def view_clusters(namespace: str = None): ) return # Exit function if not in Jupyter Notebook - from ...ray.cluster.cluster import get_current_namespace - if not namespace: namespace = get_current_namespace() diff --git a/src/codeflare_sdk/ray/__init__.py b/src/codeflare_sdk/ray/__init__.py index ab55cc82..7bd0b2c8 100644 --- a/src/codeflare_sdk/ray/__init__.py +++ b/src/codeflare_sdk/ray/__init__.py @@ -4,6 +4,14 @@ RayJobClient, ) +from .rayjobs import ( + RayJob, + ManagedClusterConfig, + RayJobDeploymentStatus, + CodeflareRayJobStatus, + RayJobInfo, +) + from .cluster import ( Cluster, ClusterConfiguration, diff --git a/src/codeflare_sdk/ray/cluster/build_ray_cluster.py b/src/codeflare_sdk/ray/cluster/build_ray_cluster.py index e8b68919..6a3984b1 100644 --- a/src/codeflare_sdk/ray/cluster/build_ray_cluster.py +++ b/src/codeflare_sdk/ray/cluster/build_ray_cluster.py @@ -21,7 +21,7 @@ from ...common.kubernetes_cluster import get_api_client, config_check from kubernetes.client.exceptions import ApiException from ...common.utils.constants import RAY_VERSION -from ...common.utils.utils import get_ray_image_for_python_version +from ...common.utils.utils import update_image import codeflare_sdk import os @@ -96,9 +96,6 @@ ), ] -# Use centralized mapping from constants (so that we only have to update constants.py) -SUPPORTED_PYTHON_VERSIONS = constants.SUPPORTED_PYTHON_VERSIONS - # RayCluster/AppWrapper builder function def build_ray_cluster(cluster: "codeflare_sdk.ray.cluster.Cluster"): @@ -245,6 +242,7 @@ def get_labels(cluster: "codeflare_sdk.ray.cluster.Cluster"): """ labels = { "controller-tools.k8s.io": "1.0", + "ray.io/cluster": cluster.config.name, # Enforced label always present } if cluster.config.labels != {}: labels.update(cluster.config.labels) @@ -269,17 +267,6 @@ def with_nb_annotations(annotations: dict): # Head/Worker container related functions -def update_image(image) -> str: - """ - The update_image() function automatically sets the image config parameter to a preset image based on Python version if not specified. - This now points to the centralized function in utils.py. - """ - if not image: - # Pull the image based on the matching Python version (or output a warning if not supported) - image = get_ray_image_for_python_version(warn_on_unsupported=True) - return image - - def get_pod_spec( cluster: "codeflare_sdk.ray.cluster.Cluster", containers: List, diff --git a/src/codeflare_sdk/ray/cluster/cluster.py b/src/codeflare_sdk/ray/cluster/cluster.py index 31b80418..cd188816 100644 --- a/src/codeflare_sdk/ray/cluster/cluster.py +++ b/src/codeflare_sdk/ray/cluster/cluster.py @@ -20,8 +20,14 @@ from time import sleep from typing import List, Optional, Tuple, Dict +import copy -from ray.job_submission import JobSubmissionClient +from ray.job_submission import JobSubmissionClient, JobStatus +import time +import uuid +import warnings + +from ...common.utils import get_current_namespace from ...common.kubernetes_cluster.auth import ( config_check, @@ -57,7 +63,6 @@ from kubernetes.client.rest import ApiException from kubernetes.client.rest import ApiException -import warnings CF_SDK_FIELD_MANAGER = "codeflare-sdk" @@ -203,6 +208,10 @@ def apply(self, force=False): self._throw_for_no_raycluster() namespace = self.config.namespace name = self.config.name + + # Regenerate resource_yaml to reflect any configuration changes + self.resource_yaml = self.create_resource() + try: self.config_check() api_instance = client.CustomObjectsApi(get_api_client()) @@ -384,9 +393,14 @@ def is_dashboard_ready(self) -> bool: bool: True if the dashboard is ready, False otherwise. """ + + dashboard_uri = self.cluster_dashboard_uri() + if dashboard_uri is None: + return False + try: response = requests.get( - self.cluster_dashboard_uri(), + dashboard_uri, headers=self._client_headers, timeout=5, verify=self._client_verify_tls, @@ -394,6 +408,10 @@ def is_dashboard_ready(self) -> bool: except requests.exceptions.SSLError: # pragma no cover # SSL exception occurs when oauth ingress has been created but cluster is not up return False + except Exception: # pragma no cover + # Any other exception (connection errors, timeouts, etc.) + return False + if response.status_code == 200: return True else: @@ -501,6 +519,8 @@ def cluster_dashboard_uri(self) -> str: ): protocol = "https" if route["spec"].get("tls") else "http" return f"{protocol}://{route['spec']['host']}" + # No route found for this cluster + return "Dashboard not available yet, have you run cluster.apply()?" else: try: api_instance = client.NetworkingV1Api(get_api_client()) @@ -637,32 +657,6 @@ def list_all_queued( return resources -def get_current_namespace(): # pragma: no cover - """ - Retrieves the current Kubernetes namespace. - - Returns: - str: - The current namespace or None if not found. - """ - if os.path.isfile("/var/run/secrets/kubernetes.io/serviceaccount/namespace"): - try: - file = open("/var/run/secrets/kubernetes.io/serviceaccount/namespace", "r") - active_context = file.readline().strip("\n") - return active_context - except Exception as e: - print("Unable to find current namespace") - print("trying to gather from current context") - try: - _, active_context = config.list_kube_config_contexts(config_check()) - except Exception as e: - return _kube_api_error_handling(e) - try: - return active_context["context"]["namespace"] - except KeyError: - return None - - def get_cluster( cluster_name: str, namespace: str = "default", @@ -762,6 +756,7 @@ def get_cluster( head_extended_resource_requests=head_extended_resources, worker_extended_resource_requests=worker_extended_resources, ) + # Ignore the warning here for the lack of a ClusterConfiguration with warnings.catch_warnings(): warnings.filterwarnings( @@ -805,6 +800,7 @@ def remove_autogenerated_fields(resource): del resource[key] else: remove_autogenerated_fields(resource[key]) + elif isinstance(resource, list): for item in resource: remove_autogenerated_fields(item) diff --git a/src/codeflare_sdk/ray/cluster/test_build_ray_cluster.py b/src/codeflare_sdk/ray/cluster/test_build_ray_cluster.py index f970d945..3a7947d3 100644 --- a/src/codeflare_sdk/ray/cluster/test_build_ray_cluster.py +++ b/src/codeflare_sdk/ray/cluster/test_build_ray_cluster.py @@ -40,7 +40,7 @@ def test_gen_names_without_name(mocker): def test_update_image_without_supported_python_version(mocker): # Mock SUPPORTED_PYTHON_VERSIONS mocker.patch.dict( - "codeflare_sdk.ray.cluster.build_ray_cluster.SUPPORTED_PYTHON_VERSIONS", + "codeflare_sdk.common.utils.constants.SUPPORTED_PYTHON_VERSIONS", { "3.11": "ray-py3.11", "3.12": "ray-py3.12", diff --git a/src/codeflare_sdk/ray/cluster/test_cluster.py b/src/codeflare_sdk/ray/cluster/test_cluster.py index 6475f7a8..164b3a81 100644 --- a/src/codeflare_sdk/ray/cluster/test_cluster.py +++ b/src/codeflare_sdk/ray/cluster/test_cluster.py @@ -956,5 +956,11 @@ def test_cluster_namespace_type_error(mocker): # Make sure to always keep this function last def test_cleanup(): - os.remove(f"{aw_dir}test-all-params.yaml") - os.remove(f"{aw_dir}aw-all-params.yaml") + # Remove files only if they exist + test_file = f"{aw_dir}test-all-params.yaml" + if os.path.exists(test_file): + os.remove(test_file) + + aw_file = f"{aw_dir}aw-all-params.yaml" + if os.path.exists(aw_file): + os.remove(aw_file) diff --git a/src/codeflare_sdk/ray/rayjobs/__init__.py b/src/codeflare_sdk/ray/rayjobs/__init__.py new file mode 100644 index 00000000..cd6b4123 --- /dev/null +++ b/src/codeflare_sdk/ray/rayjobs/__init__.py @@ -0,0 +1,3 @@ +from .rayjob import RayJob, ManagedClusterConfig +from .status import RayJobDeploymentStatus, CodeflareRayJobStatus, RayJobInfo +from .config import ManagedClusterConfig diff --git a/src/codeflare_sdk/ray/rayjobs/config.py b/src/codeflare_sdk/ray/rayjobs/config.py new file mode 100644 index 00000000..5b724272 --- /dev/null +++ b/src/codeflare_sdk/ray/rayjobs/config.py @@ -0,0 +1,502 @@ +# Copyright 2022 IBM, Red Hat +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +""" +The config sub-module contains the definition of the ManagedClusterConfig dataclass, +which is used to specify resource requirements and other details when creating a +Cluster object. +""" + +import pathlib +from dataclasses import dataclass, field, fields +from typing import Dict, List, Optional, Union, get_args, get_origin, Any, Tuple +from kubernetes.client import ( + V1LocalObjectReference, + V1SecretVolumeSource, + V1Toleration, + V1Volume, + V1VolumeMount, + V1ObjectMeta, + V1Container, + V1ContainerPort, + V1Lifecycle, + V1ExecAction, + V1LifecycleHandler, + V1EnvVar, + V1PodTemplateSpec, + V1PodSpec, + V1ResourceRequirements, +) + +import logging + +from ...common.utils.constants import MOUNT_PATH, RAY_VERSION +from ...common.utils.utils import update_image + +logger = logging.getLogger(__name__) + +dir = pathlib.Path(__file__).parent.parent.resolve() + +# https://docs.ray.io/en/latest/ray-core/scheduling/accelerators.html +DEFAULT_ACCELERATORS = { + "nvidia.com/gpu": "GPU", + "intel.com/gpu": "GPU", + "amd.com/gpu": "GPU", + "aws.amazon.com/neuroncore": "neuron_cores", + "google.com/tpu": "TPU", + "habana.ai/gaudi": "HPU", + "huawei.com/Ascend910": "NPU", + "huawei.com/Ascend310": "NPU", +} + + +@dataclass +class ManagedClusterConfig: + """ + This dataclass is used to specify resource requirements and other details for RayJobs. + The cluster name and namespace are automatically derived from the RayJob configuration. + + Args: + head_accelerators: + A dictionary of extended resource requests for the head node. ex: {"nvidia.com/gpu": 1} + head_tolerations: + List of tolerations for head nodes. + num_workers: + The number of workers to create. + worker_tolerations: + List of tolerations for worker nodes. + envs: + A dictionary of environment variables to set for the cluster. + image: + The image to use for the cluster. + image_pull_secrets: + A list of image pull secrets to use for the cluster. + labels: + A dictionary of labels to apply to the cluster. + worker_accelerators: + A dictionary of extended resource requests for each worker. ex: {"nvidia.com/gpu": 1} + accelerator_configs: + A dictionary of custom resource mappings to map extended resource requests to RayCluster resource names. + Defaults to DEFAULT_ACCELERATORS but can be overridden with custom mappings. + annotations: + A dictionary of annotations to apply to the Job. + volumes: + A list of V1Volume objects to add to the Cluster + volume_mounts: + A list of V1VolumeMount objects to add to the Cluster + """ + + head_cpu_requests: Union[int, str] = 2 + head_cpu_limits: Union[int, str] = 2 + head_memory_requests: Union[int, str] = 8 + head_memory_limits: Union[int, str] = 8 + head_accelerators: Dict[str, Union[str, int]] = field(default_factory=dict) + head_tolerations: Optional[List[V1Toleration]] = None + worker_cpu_requests: Union[int, str] = 1 + worker_cpu_limits: Union[int, str] = 1 + num_workers: int = 1 + worker_memory_requests: Union[int, str] = 2 + worker_memory_limits: Union[int, str] = 2 + worker_tolerations: Optional[List[V1Toleration]] = None + envs: Dict[str, str] = field(default_factory=dict) + image: str = "" + image_pull_secrets: List[str] = field(default_factory=list) + labels: Dict[str, str] = field(default_factory=dict) + worker_accelerators: Dict[str, Union[str, int]] = field(default_factory=dict) + accelerator_configs: Dict[str, str] = field( + default_factory=lambda: DEFAULT_ACCELERATORS.copy() + ) + annotations: Dict[str, str] = field(default_factory=dict) + volumes: list[V1Volume] = field(default_factory=list) + volume_mounts: list[V1VolumeMount] = field(default_factory=list) + + def __post_init__(self): + self.envs["RAY_USAGE_STATS_ENABLED"] = "0" + + self._validate_types() + self._memory_to_string() + self._validate_gpu_config(self.head_accelerators) + self._validate_gpu_config(self.worker_accelerators) + + def _validate_gpu_config(self, gpu_config: Dict[str, int]): + for k in gpu_config.keys(): + if k not in self.accelerator_configs.keys(): + raise ValueError( + f"GPU configuration '{k}' not found in accelerator_configs, available resources are {list(self.accelerator_configs.keys())}, to add more supported resources use accelerator_configs. i.e. accelerator_configs = {{'{k}': 'FOO_BAR'}}" + ) + + def _memory_to_string(self): + if isinstance(self.head_memory_requests, int): + self.head_memory_requests = f"{self.head_memory_requests}G" + if isinstance(self.head_memory_limits, int): + self.head_memory_limits = f"{self.head_memory_limits}G" + if isinstance(self.worker_memory_requests, int): + self.worker_memory_requests = f"{self.worker_memory_requests}G" + if isinstance(self.worker_memory_limits, int): + self.worker_memory_limits = f"{self.worker_memory_limits}G" + + def _validate_types(self): + """Validate the types of all fields in the ManagedClusterConfig dataclass.""" + errors = [] + for field_info in fields(self): + value = getattr(self, field_info.name) + expected_type = field_info.type + if not self._is_type(value, expected_type): + errors.append(f"'{field_info.name}' should be of type {expected_type}.") + + if errors: + raise TypeError("Type validation failed:\n" + "\n".join(errors)) + + @staticmethod + def _is_type(value, expected_type): + """Check if the value matches the expected type.""" + + def check_type(value, expected_type): + origin_type = get_origin(expected_type) + args = get_args(expected_type) + if origin_type is Union: + return any(check_type(value, union_type) for union_type in args) + if origin_type is list: + if value is not None: + return all(check_type(elem, args[0]) for elem in (value or [])) + else: + return True + if origin_type is dict: + if value is not None: + return all( + check_type(k, args[0]) and check_type(v, args[1]) + for k, v in value.items() + ) + else: + return True + if origin_type is tuple: + return all(check_type(elem, etype) for elem, etype in zip(value, args)) + if expected_type is int: + return isinstance(value, int) and not isinstance(value, bool) + if expected_type is bool: + return isinstance(value, bool) + return isinstance(value, expected_type) + + return check_type(value, expected_type) + + def build_ray_cluster_spec(self, cluster_name: str) -> Dict[str, Any]: + """ + Build the RayCluster spec from ManagedClusterConfig for embedding in RayJob. + + Args: + self: The cluster configuration object (ManagedClusterConfig) + cluster_name: The name for the cluster (derived from RayJob name) + + Returns: + Dict containing the RayCluster spec for embedding in RayJob + """ + ray_cluster_spec = { + "rayVersion": RAY_VERSION, + "enableInTreeAutoscaling": False, # Required for Kueue-managed jobs + "headGroupSpec": self._build_head_group_spec(), + "workerGroupSpecs": [self._build_worker_group_spec(cluster_name)], + } + + return ray_cluster_spec + + def _build_head_group_spec(self) -> Dict[str, Any]: + """Build the head group specification.""" + return { + "serviceType": "ClusterIP", + "enableIngress": False, + "rayStartParams": self._build_head_ray_params(), + "template": V1PodTemplateSpec( + metadata=V1ObjectMeta(annotations=self.annotations), + spec=self._build_pod_spec(self._build_head_container(), is_head=True), + ), + } + + def _build_worker_group_spec(self, cluster_name: str) -> Dict[str, Any]: + """Build the worker group specification.""" + return { + "replicas": self.num_workers, + "minReplicas": self.num_workers, + "maxReplicas": self.num_workers, + "groupName": f"worker-group-{cluster_name}", + "rayStartParams": self._build_worker_ray_params(), + "template": V1PodTemplateSpec( + metadata=V1ObjectMeta(annotations=self.annotations), + spec=self._build_pod_spec( + self._build_worker_container(), + is_head=False, + ), + ), + } + + def _build_head_ray_params(self) -> Dict[str, str]: + """Build Ray start parameters for head node.""" + params = { + "dashboard-host": "0.0.0.0", + "block": "true", + } + + # Add GPU count if specified + if self.head_accelerators: + gpu_count = sum( + count + for resource_type, count in self.head_accelerators.items() + if "gpu" in resource_type.lower() + ) + if gpu_count > 0: + params["num-gpus"] = str(gpu_count) + + return params + + def _build_worker_ray_params(self) -> Dict[str, str]: + """Build Ray start parameters for worker nodes.""" + params = { + "block": "true", + } + + # Add GPU count if specified + if self.worker_accelerators: + gpu_count = sum( + count + for resource_type, count in self.worker_accelerators.items() + if "gpu" in resource_type.lower() + ) + if gpu_count > 0: + params["num-gpus"] = str(gpu_count) + + return params + + def _build_head_container(self) -> V1Container: + """Build the head container specification.""" + container = V1Container( + name="ray-head", + image=update_image(self.image), + image_pull_policy="IfNotPresent", # Always IfNotPresent for RayJobs + ports=[ + V1ContainerPort(name="gcs", container_port=6379), + V1ContainerPort(name="dashboard", container_port=8265), + V1ContainerPort(name="client", container_port=10001), + ], + lifecycle=V1Lifecycle( + pre_stop=V1LifecycleHandler( + _exec=V1ExecAction(command=["/bin/sh", "-c", "ray stop"]) + ) + ), + resources=self._build_resource_requirements( + self.head_cpu_requests, + self.head_cpu_limits, + self.head_memory_requests, + self.head_memory_limits, + self.head_accelerators, + ), + volume_mounts=self._generate_volume_mounts(), + env=self._build_env_vars() if hasattr(self, "envs") and self.envs else None, + ) + + return container + + def _build_worker_container(self) -> V1Container: + """Build the worker container specification.""" + container = V1Container( + name="ray-worker", + image=update_image(self.image), + image_pull_policy="IfNotPresent", # Always IfNotPresent for RayJobs + lifecycle=V1Lifecycle( + pre_stop=V1LifecycleHandler( + _exec=V1ExecAction(command=["/bin/sh", "-c", "ray stop"]) + ) + ), + resources=self._build_resource_requirements( + self.worker_cpu_requests, + self.worker_cpu_limits, + self.worker_memory_requests, + self.worker_memory_limits, + self.worker_accelerators, + ), + volume_mounts=self._generate_volume_mounts(), + env=self._build_env_vars() if hasattr(self, "envs") and self.envs else None, + ) + + return container + + def _build_resource_requirements( + self, + cpu_requests: Union[int, str], + cpu_limits: Union[int, str], + memory_requests: Union[int, str], + memory_limits: Union[int, str], + extended_resource_requests: Dict[str, Union[int, str]] = None, + ) -> V1ResourceRequirements: + """Build Kubernetes resource requirements.""" + resource_requirements = V1ResourceRequirements( + requests={"cpu": cpu_requests, "memory": memory_requests}, + limits={"cpu": cpu_limits, "memory": memory_limits}, + ) + + # Add extended resources (e.g., GPUs) + if extended_resource_requests: + for resource_type, amount in extended_resource_requests.items(): + resource_requirements.limits[resource_type] = amount + resource_requirements.requests[resource_type] = amount + + return resource_requirements + + def _build_pod_spec(self, container: V1Container, is_head: bool) -> V1PodSpec: + """Build the pod specification.""" + pod_spec = V1PodSpec( + containers=[container], + volumes=self._generate_volumes(), + restart_policy="Never", # RayJobs should not restart + ) + + # Add tolerations if specified + if is_head and hasattr(self, "head_tolerations") and self.head_tolerations: + pod_spec.tolerations = self.head_tolerations + elif ( + not is_head + and hasattr(self, "worker_tolerations") + and self.worker_tolerations + ): + pod_spec.tolerations = self.worker_tolerations + + # Add image pull secrets if specified + if hasattr(self, "image_pull_secrets") and self.image_pull_secrets: + pod_spec.image_pull_secrets = [ + V1LocalObjectReference(name=secret) + for secret in self.image_pull_secrets + ] + + return pod_spec + + def _generate_volume_mounts(self) -> list: + """Generate volume mounts for the container.""" + volume_mounts = [] + + # Add custom volume mounts if specified + if hasattr(self, "volume_mounts") and self.volume_mounts: + volume_mounts.extend(self.volume_mounts) + + return volume_mounts + + def _generate_volumes(self) -> list: + """Generate volumes for the pod.""" + volumes = [] + + # Add custom volumes if specified + if hasattr(self, "volumes") and self.volumes: + volumes.extend(self.volumes) + + return volumes + + def _build_env_vars(self) -> list: + """Build environment variables list.""" + return [V1EnvVar(name=key, value=value) for key, value in self.envs.items()] + + def add_file_volumes(self, secret_name: str, mount_path: str = MOUNT_PATH): + """ + Add file volume and mount references to cluster configuration. + + Args: + secret_name: Name of the Secret containing files + mount_path: Where to mount files in containers (default: /home/ray/scripts) + """ + # Check if file volume already exists + volume_name = "ray-job-files" + existing_volume = next( + (v for v in self.volumes if getattr(v, "name", None) == volume_name), None + ) + if existing_volume: + logger.debug(f"File volume '{volume_name}' already exists, skipping...") + return + + # Check if file mount already exists + existing_mount = next( + (m for m in self.volume_mounts if getattr(m, "name", None) == volume_name), + None, + ) + if existing_mount: + logger.debug( + f"File volume mount '{volume_name}' already exists, skipping..." + ) + return + + # Add file volume to cluster configuration + file_volume = V1Volume( + name=volume_name, secret=V1SecretVolumeSource(secret_name=secret_name) + ) + self.volumes.append(file_volume) + + # Add file volume mount to cluster configuration + file_mount = V1VolumeMount(name=volume_name, mount_path=mount_path) + self.volume_mounts.append(file_mount) + + logger.info( + f"Added file volume '{secret_name}' to cluster config: mount_path={mount_path}" + ) + + def validate_secret_size(self, files: Dict[str, str]) -> None: + total_size = sum(len(content.encode("utf-8")) for content in files.values()) + if total_size > 1024 * 1024: # 1MB + raise ValueError( + f"Secret size exceeds 1MB limit. Total size: {total_size} bytes" + ) + + def build_file_secret_spec( + self, job_name: str, namespace: str, files: Dict[str, str] + ) -> Dict[str, Any]: + """ + Build Secret specification for files + + Args: + job_name: Name of the RayJob (used for Secret naming) + namespace: Kubernetes namespace + files: Dictionary of file_name -> file_content + + Returns: + Dict: Secret specification ready for Kubernetes API + """ + secret_name = f"{job_name}-files" + return { + "apiVersion": "v1", + "kind": "Secret", + "type": "Opaque", + "metadata": { + "name": secret_name, + "namespace": namespace, + "labels": { + "ray.io/job-name": job_name, + "app.kubernetes.io/managed-by": "codeflare-sdk", + "app.kubernetes.io/component": "rayjob-files", + }, + }, + "data": files, + } + + def build_file_volume_specs( + self, secret_name: str, mount_path: str = MOUNT_PATH + ) -> Tuple[Dict[str, Any], Dict[str, Any]]: + """ + Build volume and mount specifications for files + + Args: + secret_name: Name of the Secret containing files + mount_path: Where to mount files in containers + + Returns: + Tuple of (volume_spec, mount_spec) as dictionaries + """ + volume_spec = {"name": "ray-job-files", "secret": {"secretName": secret_name}} + + mount_spec = {"name": "ray-job-files", "mountPath": mount_path} + + return volume_spec, mount_spec diff --git a/src/codeflare_sdk/ray/rayjobs/pretty_print.py b/src/codeflare_sdk/ray/rayjobs/pretty_print.py new file mode 100644 index 00000000..34e8dfa1 --- /dev/null +++ b/src/codeflare_sdk/ray/rayjobs/pretty_print.py @@ -0,0 +1,117 @@ +# Copyright 2025 IBM, Red Hat +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +""" +This sub-module exists primarily to be used internally by the RayJob object +(in the rayjob sub-module) for pretty-printing job status and details. +""" + +from rich.console import Console +from rich.table import Table +from rich.panel import Panel +from typing import Tuple, Optional + +from .status import RayJobDeploymentStatus, RayJobInfo + + +def print_job_status(job_info: RayJobInfo): + """ + Pretty print the job status in a format similar to cluster status. + """ + status_display, header_color = _get_status_display(job_info.status) + + # Create main info table + table = _create_info_table(header_color, job_info.name, status_display) + table.add_row(f"[bold]Job ID:[/bold] {job_info.job_id}") + table.add_row(f"[bold]Status:[/bold] {job_info.status.value}") + table.add_row(f"[bold]RayCluster:[/bold] {job_info.cluster_name}") + table.add_row(f"[bold]Namespace:[/bold] {job_info.namespace}") + + # Add timing information if available + if job_info.start_time: + table.add_row() + table.add_row(f"[bold]Started:[/bold] {job_info.start_time}") + + # Add attempt counts if there are failures + if job_info.failed_attempts > 0: + table.add_row(f"[bold]Failed Attempts:[/bold] {job_info.failed_attempts}") + + _print_table_in_panel(table) + + +def print_no_job_found(job_name: str, namespace: str): + """ + Print a message when no job is found. + """ + # Create table with error message + table = _create_info_table( + "[white on red][bold]Name", job_name, "[bold red]No RayJob found" + ) + table.add_row() + table.add_row("Please run rayjob.submit() to submit a job.") + table.add_row() + table.add_row(f"[bold]Namespace:[/bold] {namespace}") + + _print_table_in_panel(table) + + +def _get_status_display(status: RayJobDeploymentStatus) -> Tuple[str, str]: + """ + Get the display string and header color for a given status. + + Returns: + Tuple of (status_display, header_color) + """ + status_mapping = { + RayJobDeploymentStatus.COMPLETE: ( + "Complete :white_heavy_check_mark:", + "[white on green][bold]Name", + ), + RayJobDeploymentStatus.RUNNING: ("Running :gear:", "[white on blue][bold]Name"), + RayJobDeploymentStatus.FAILED: ("Failed :x:", "[white on red][bold]Name"), + RayJobDeploymentStatus.SUSPENDED: ( + "Suspended :pause_button:", + "[white on yellow][bold]Name", + ), + } + + return status_mapping.get( + status, ("Unknown :question:", "[white on red][bold]Name") + ) + + +def _create_info_table(header_color: str, name: str, status_display: str) -> Table: + """ + Create a standardized info table with header and status. + + Returns: + Table with header row, name/status row, and empty separator row + """ + table = Table(box=None, show_header=False) + table.add_row(header_color) + table.add_row("[bold underline]" + name, status_display) + table.add_row() # Empty separator row + return table + + +def _print_table_in_panel(table: Table): + """ + Print a table wrapped in a consistent panel format. + """ + console = Console() + main_table = Table( + box=None, title="[bold] :package: CodeFlare RayJob Status :package:" + ) + main_table.add_row(Panel.fit(table)) + console.print(main_table) diff --git a/src/codeflare_sdk/ray/rayjobs/rayjob.py b/src/codeflare_sdk/ray/rayjobs/rayjob.py new file mode 100644 index 00000000..65478087 --- /dev/null +++ b/src/codeflare_sdk/ray/rayjobs/rayjob.py @@ -0,0 +1,615 @@ +# Copyright 2022-2025 IBM, Red Hat +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +""" +RayJob client for submitting and managing Ray jobs using the kuberay python client. +""" + +import logging +import os +import re +import warnings +from typing import Dict, Any, Optional, Tuple, Union + +from ray.runtime_env import RuntimeEnv +from codeflare_sdk.common.kueue.kueue import get_default_kueue_name +from codeflare_sdk.common.utils.constants import MOUNT_PATH + +from codeflare_sdk.common.utils.utils import get_ray_image_for_python_version +from python_client.kuberay_job_api import RayjobApi +from python_client.kuberay_cluster_api import RayClusterApi +from codeflare_sdk.ray.rayjobs.config import ManagedClusterConfig +from codeflare_sdk.ray.rayjobs.runtime_env import ( + create_file_secret, + extract_all_local_files, + process_runtime_env, +) + +from ...common.utils import get_current_namespace +from ...common.utils.validation import validate_ray_version_compatibility + +from .status import ( + RayJobDeploymentStatus, + CodeflareRayJobStatus, + RayJobInfo, +) +from . import pretty_print + + +logger = logging.getLogger(__name__) + + +class RayJob: + """ + A client for managing Ray jobs using the KubeRay operator. + + This class provides a simplified interface for submitting and managing + RayJob CRs (using the KubeRay RayJob python client). + """ + + def __init__( + self, + job_name: str, + entrypoint: str, + cluster_name: Optional[str] = None, + cluster_config: Optional[ManagedClusterConfig] = None, + namespace: Optional[str] = None, + runtime_env: Optional[Union[RuntimeEnv, Dict[str, Any]]] = None, + ttl_seconds_after_finished: int = 0, + active_deadline_seconds: Optional[int] = None, + local_queue: Optional[str] = None, + ): + """ + Initialize a RayJob instance. + + Args: + job_name: The name for the Ray job + entrypoint: The Python script or command to run (required) + cluster_name: The name of an existing Ray cluster (optional if cluster_config provided) + cluster_config: Configuration for creating a new cluster (optional if cluster_name provided) + namespace: The Kubernetes namespace (auto-detected if not specified) + runtime_env: Ray runtime environment configuration. Can be: + - RuntimeEnv object from ray.runtime_env + - Dict with keys like 'working_dir', 'pip', 'env_vars', etc. + Example: {"working_dir": "./my-scripts", "pip": ["requests"]} + ttl_seconds_after_finished: Seconds to wait before cleanup after job finishes (default: 0) + active_deadline_seconds: Maximum time the job can run before being terminated (optional) + local_queue: The Kueue LocalQueue to submit the job to (optional) + + Note: + - True if cluster_config is provided (new cluster will be cleaned up) + - False if cluster_name is provided (existing cluster will not be shut down) + - User can explicitly set this value to override auto-detection + """ + if cluster_name is None and cluster_config is None: + raise ValueError( + "❌ Configuration Error: You must provide either 'cluster_name' (for existing cluster) " + "or 'cluster_config' (to create new cluster), but not both." + ) + + if cluster_name is not None and cluster_config is not None: + raise ValueError( + "❌ Configuration Error: You cannot specify both 'cluster_name' and 'cluster_config'. " + "Choose one approach:\n" + "• Use 'cluster_name' to connect to an existing cluster\n" + "• Use 'cluster_config' to create a new cluster" + ) + + if cluster_config is None and cluster_name is None: + raise ValueError( + "❌ Configuration Error: When not providing 'cluster_config', 'cluster_name' is required " + "to specify which existing cluster to use." + ) + + self.name = job_name + self.entrypoint = entrypoint + + # Convert dict to RuntimeEnv if needed for user convenience + if isinstance(runtime_env, dict): + self.runtime_env = RuntimeEnv(**runtime_env) + else: + self.runtime_env = runtime_env + + self.ttl_seconds_after_finished = ttl_seconds_after_finished + self.active_deadline_seconds = active_deadline_seconds + self.local_queue = local_queue + + if namespace is None: + detected_namespace = get_current_namespace() + if detected_namespace: + self.namespace = detected_namespace + logger.info(f"Auto-detected namespace: {self.namespace}") + else: + raise ValueError( + "❌ Configuration Error: Could not auto-detect Kubernetes namespace. " + "Please explicitly specify the 'namespace' parameter. " + ) + else: + self.namespace = namespace + + self._cluster_name = cluster_name + self._cluster_config = cluster_config + + if cluster_config is not None: + self.cluster_name = f"{job_name}-cluster" + logger.info(f"Creating new cluster: {self.cluster_name}") + else: + # Using existing cluster: cluster_name must be provided + if cluster_name is None: + raise ValueError( + "❌ Configuration Error: a 'cluster_name' is required when not providing 'cluster_config'" + ) + self.cluster_name = cluster_name + logger.info(f"Using existing cluster: {self.cluster_name}") + + self._api = RayjobApi() + self._cluster_api = RayClusterApi() + + logger.info(f"Initialized RayJob: {self.name} in namespace: {self.namespace}") + + def submit(self) -> str: + if not self.entrypoint: + raise ValueError("Entrypoint must be provided to submit a RayJob") + + # Validate configuration before submitting + self._validate_ray_version_compatibility() + self._validate_working_dir_entrypoint() + + # Extract files from entrypoint and runtime_env working_dir + files = extract_all_local_files(self) + + rayjob_cr = self._build_rayjob_cr() + + logger.info(f"Submitting RayJob {self.name} to Kuberay operator") + result = self._api.submit_job(k8s_namespace=self.namespace, job=rayjob_cr) + + if result: + logger.info(f"Successfully submitted RayJob {self.name}") + + # Create Secret with owner reference after RayJob exists + if files: + create_file_secret(self, files, result) + + return self.name + else: + raise RuntimeError(f"Failed to submit RayJob {self.name}") + + def stop(self): + """ + Suspend the Ray job. + """ + stopped = self._api.suspend_job(name=self.name, k8s_namespace=self.namespace) + if stopped: + logger.info(f"Successfully stopped the RayJob {self.name}") + return True + else: + raise RuntimeError(f"Failed to stop the RayJob {self.name}") + + def resubmit(self): + """ + Resubmit the Ray job. + """ + if self._api.resubmit_job(name=self.name, k8s_namespace=self.namespace): + logger.info(f"Successfully resubmitted the RayJob {self.name}") + return True + else: + raise RuntimeError(f"Failed to resubmit the RayJob {self.name}") + + def delete(self): + """ + Delete the Ray job. + Returns True if deleted successfully or if already deleted. + """ + deleted = self._api.delete_job(name=self.name, k8s_namespace=self.namespace) + if deleted: + logger.info(f"Successfully deleted the RayJob {self.name}") + return True + else: + # The python client logs "rayjob custom resource already deleted" + # and returns False when the job doesn't exist. + # This is not an error - treat it as successful deletion. + logger.info(f"RayJob {self.name} already deleted or does not exist") + return True + + def _build_rayjob_cr(self) -> Dict[str, Any]: + """ + Build the RayJob custom resource specification using native RayJob capabilities. + """ + rayjob_cr = { + "apiVersion": "ray.io/v1", + "kind": "RayJob", + "metadata": { + "name": self.name, + "namespace": self.namespace, + }, + "spec": { + "entrypoint": self.entrypoint, + "ttlSecondsAfterFinished": self.ttl_seconds_after_finished, + "shutdownAfterJobFinishes": self._cluster_config is not None, + }, + } + + # Extract files once and use for both runtime_env and submitter pod + files = extract_all_local_files(self) + + labels = {} + # If cluster_config is provided, use the local_queue from the cluster_config + if self._cluster_config is not None: + if self.local_queue: + labels["kueue.x-k8s.io/queue-name"] = self.local_queue + else: + default_queue = get_default_kueue_name(self.namespace) + if default_queue: + labels["kueue.x-k8s.io/queue-name"] = default_queue + else: + # No default queue found, use "default" as fallback + labels["kueue.x-k8s.io/queue-name"] = "default" + logger.warning( + f"No default Kueue LocalQueue found in namespace '{self.namespace}'. " + f"Using 'default' as the queue name. If a LocalQueue named 'default' " + f"does not exist, the RayJob submission will fail. " + f"To fix this, please explicitly specify the 'local_queue' parameter." + ) + + rayjob_cr["metadata"]["labels"] = labels + + # When using Kueue (queue label present), start with suspend=true + # Kueue will unsuspend the job once the workload is admitted + if labels.get("kueue.x-k8s.io/queue-name"): + rayjob_cr["spec"]["suspend"] = True + + # Add active deadline if specified + if self.active_deadline_seconds: + rayjob_cr["spec"]["activeDeadlineSeconds"] = self.active_deadline_seconds + + # Add runtime environment (can be inferred even if not explicitly specified) + processed_runtime_env = process_runtime_env(self, files) + if processed_runtime_env: + rayjob_cr["spec"]["runtimeEnvYAML"] = processed_runtime_env + + # Add submitterPodTemplate if we have files to mount + if files: + secret_name = f"{self.name}-files" + rayjob_cr["spec"][ + "submitterPodTemplate" + ] = self._build_submitter_pod_template(files, secret_name) + + # Configure cluster: either use existing or create new + if self._cluster_config is not None: + ray_cluster_spec = self._cluster_config.build_ray_cluster_spec( + cluster_name=self.cluster_name + ) + + logger.info( + f"Built RayCluster spec using RayJob-specific builder for cluster: {self.cluster_name}" + ) + + rayjob_cr["spec"]["rayClusterSpec"] = ray_cluster_spec + + logger.info(f"RayJob will create new cluster: {self.cluster_name}") + else: + # Use clusterSelector to reference existing cluster + rayjob_cr["spec"]["clusterSelector"] = {"ray.io/cluster": self.cluster_name} + logger.info(f"RayJob will use existing cluster: {self.cluster_name}") + + return rayjob_cr + + def _build_submitter_pod_template( + self, files: Dict[str, str], secret_name: str + ) -> Dict[str, Any]: + """ + Build submitterPodTemplate with Secret volume mount for local files. + + If files contain working_dir.zip, an init container will unzip it before + the main submitter container runs. + + Args: + files: Dict of file_name -> file_content + secret_name: Name of the Secret containing the files + + Returns: + submitterPodTemplate specification + """ + from codeflare_sdk.ray.rayjobs.runtime_env import UNZIP_PATH + + # Image has to be hard coded for the job submitter + image = get_ray_image_for_python_version() + if ( + self._cluster_config + and hasattr(self._cluster_config, "image") + and self._cluster_config.image + ): + image = self._cluster_config.image + + # Build Secret items for each file + secret_items = [] + entrypoint_path = files.get( + "__entrypoint_path__" + ) # Metadata for single file case + + for file_name in files.keys(): + if file_name == "__entrypoint_path__": + continue # Skip metadata key + + # For single file case, use the preserved path structure + if entrypoint_path: + secret_items.append({"key": file_name, "path": entrypoint_path}) + else: + secret_items.append({"key": file_name, "path": file_name}) + + # Check if we need to unzip working_dir + has_working_dir_zip = "working_dir.zip" in files + + # Base volume mounts for main container + volume_mounts = [{"name": "ray-job-files", "mountPath": MOUNT_PATH}] + + # If we have a zip file, we need shared volume for unzipped content + if has_working_dir_zip: + volume_mounts.append( + {"name": "unzipped-working-dir", "mountPath": UNZIP_PATH} + ) + + submitter_pod_template = { + "spec": { + "restartPolicy": "Never", + "containers": [ + { + "name": "ray-job-submitter", + "image": image, + "volumeMounts": volume_mounts, + } + ], + "volumes": [ + { + "name": "ray-job-files", + "secret": { + "secretName": secret_name, + "items": secret_items, + }, + } + ], + } + } + + # Add init container and volume for unzipping if needed + if has_working_dir_zip: + # Add emptyDir volume for unzipped content + submitter_pod_template["spec"]["volumes"].append( + {"name": "unzipped-working-dir", "emptyDir": {}} + ) + + # Add init container to unzip before KubeRay's submitter runs + submitter_pod_template["spec"]["initContainers"] = [ + { + "name": "unzip-working-dir", + "image": image, + "command": ["/bin/sh", "-c"], + "args": [ + # Decode base64 zip, save to temp file, extract, cleanup + f"mkdir -p {UNZIP_PATH} && " + f"python3 -m base64 -d {MOUNT_PATH}/working_dir.zip > /tmp/working_dir.zip && " + f"python3 -m zipfile -e /tmp/working_dir.zip {UNZIP_PATH}/ && " + f"rm /tmp/working_dir.zip && " + f"echo 'Successfully unzipped working_dir to {UNZIP_PATH}' && " + f"ls -la {UNZIP_PATH}" + ], + "volumeMounts": volume_mounts, + } + ] + logger.info(f"Added init container to unzip working_dir to {UNZIP_PATH}") + + logger.info( + f"Built submitterPodTemplate with {len(files)} files mounted at {MOUNT_PATH}, using image: {image}" + ) + return submitter_pod_template + + def _validate_ray_version_compatibility(self): + """ + Validate Ray version compatibility for cluster_config image. + Raises ValueError if there is a version mismatch. + """ + # Validate cluster_config image if creating new cluster + if self._cluster_config is not None: + self._validate_cluster_config_image() + + def _validate_cluster_config_image(self): + """ + Validate that the Ray version in cluster_config image matches the SDK's Ray version. + """ + if not hasattr(self._cluster_config, "image"): + logger.debug( + "No image attribute found in cluster config, skipping validation" + ) + return + + image = self._cluster_config.image + if not image: + logger.debug("Cluster config image is empty, skipping validation") + return + + if not isinstance(image, str): + logger.warning( + f"Cluster config image should be a string, got {type(image).__name__}: {image}" + ) + return # Skip validation for malformed image + + is_compatible, is_warning, message = validate_ray_version_compatibility(image) + if not is_compatible: + raise ValueError(f"Cluster config image: {message}") + elif is_warning: + warnings.warn(f"Cluster config image: {message}") + + def _validate_working_dir_entrypoint(self): + """ + Validate entrypoint file configuration. + + Checks: + 1. Entrypoint doesn't redundantly reference working_dir + 2. Local files exist before submission + + Raises ValueError if validation fails. + """ + # Skip validation for inline commands (python -c, etc.) + if re.search(r"\s+-c\s+", self.entrypoint): + return + + # Match Python file references only + file_pattern = r"(?:python\d?\s+)?([./\w/-]+\.py)" + matches = re.findall(file_pattern, self.entrypoint) + + if not matches: + return + + entrypoint_path = matches[0] + + # Get working_dir from runtime_env + runtime_env_dict = None + working_dir = None + + if self.runtime_env: + runtime_env_dict = ( + self.runtime_env.to_dict() + if hasattr(self.runtime_env, "to_dict") + else self.runtime_env + ) + if runtime_env_dict and "working_dir" in runtime_env_dict: + working_dir = runtime_env_dict["working_dir"] + + # Skip all validation for remote working_dir + if working_dir and not os.path.isdir(working_dir): + return + + # Case 1: Local working_dir - check redundancy and file existence + if working_dir: + normalized_working_dir = os.path.normpath(working_dir) + normalized_entrypoint = os.path.normpath(entrypoint_path) + + # Check for redundant directory reference + if normalized_entrypoint.startswith(normalized_working_dir + os.sep): + relative_to_working_dir = os.path.relpath( + normalized_entrypoint, normalized_working_dir + ) + working_dir_basename = os.path.basename(normalized_working_dir) + redundant_nested_path = os.path.join( + normalized_working_dir, + working_dir_basename, + relative_to_working_dir, + ) + + if not os.path.exists(redundant_nested_path): + raise ValueError( + f"❌ Working directory conflict detected:\n" + f" working_dir: '{working_dir}'\n" + f" entrypoint references: '{entrypoint_path}'\n" + f"\n" + f"This will fail because the entrypoint runs from within working_dir.\n" + f"It would look for: '{redundant_nested_path}' (which doesn't exist)\n" + f"\n" + f"Fix: Remove the directory prefix from your entrypoint:\n" + f' entrypoint = "python {relative_to_working_dir}"' + ) + + # Check file exists within working_dir + if not normalized_entrypoint.startswith(normalized_working_dir + os.sep): + # Use normalized_working_dir (absolute path) for proper file existence check + full_entrypoint_path = os.path.join( + normalized_working_dir, entrypoint_path + ) + if not os.path.isfile(full_entrypoint_path): + raise ValueError( + f"❌ Entrypoint file not found:\n" + f" Looking for: '{full_entrypoint_path}'\n" + f" (working_dir: '{working_dir}', entrypoint file: '{entrypoint_path}')\n" + f"\n" + f"Please ensure the file exists at the expected location." + ) + + # Case 2: No working_dir - validate local file exists + else: + if not os.path.isfile(entrypoint_path): + raise ValueError( + f"❌ Entrypoint file not found: '{entrypoint_path}'\n" + f"\n" + f"Please ensure the file exists at the specified path." + ) + + def status( + self, print_to_console: bool = True + ) -> Tuple[CodeflareRayJobStatus, bool]: + """ + Get the status of the Ray job. + + Args: + print_to_console (bool): Whether to print formatted status to console (default: True) + + Returns: + Tuple of (CodeflareRayJobStatus, ready: bool) where ready indicates job completion + """ + status_data = self._api.get_job_status( + name=self.name, k8s_namespace=self.namespace + ) + + if not status_data: + if print_to_console: + pretty_print.print_no_job_found(self.name, self.namespace) + return CodeflareRayJobStatus.UNKNOWN, False + + # Map deployment status to our enums + deployment_status_str = status_data.get("jobDeploymentStatus", "Unknown") + + try: + deployment_status = RayJobDeploymentStatus(deployment_status_str) + except ValueError: + deployment_status = RayJobDeploymentStatus.UNKNOWN + + # Create RayJobInfo dataclass + job_info = RayJobInfo( + name=self.name, + job_id=status_data.get("jobId", ""), + status=deployment_status, + namespace=self.namespace, + cluster_name=self.cluster_name, + start_time=status_data.get("startTime"), + end_time=status_data.get("endTime"), + failed_attempts=status_data.get("failed", 0), + succeeded_attempts=status_data.get("succeeded", 0), + ) + + # Map to CodeFlare status and determine readiness + codeflare_status, ready = self._map_to_codeflare_status(deployment_status) + + if print_to_console: + pretty_print.print_job_status(job_info) + + return codeflare_status, ready + + def _map_to_codeflare_status( + self, deployment_status: RayJobDeploymentStatus + ) -> Tuple[CodeflareRayJobStatus, bool]: + """ + Map deployment status to CodeFlare status and determine readiness. + + Returns: + Tuple of (CodeflareRayJobStatus, ready: bool) + """ + status_mapping = { + RayJobDeploymentStatus.COMPLETE: (CodeflareRayJobStatus.COMPLETE, True), + RayJobDeploymentStatus.RUNNING: (CodeflareRayJobStatus.RUNNING, False), + RayJobDeploymentStatus.FAILED: (CodeflareRayJobStatus.FAILED, False), + RayJobDeploymentStatus.SUSPENDED: (CodeflareRayJobStatus.SUSPENDED, False), + } + + return status_mapping.get( + deployment_status, (CodeflareRayJobStatus.UNKNOWN, False) + ) diff --git a/src/codeflare_sdk/ray/rayjobs/runtime_env.py b/src/codeflare_sdk/ray/rayjobs/runtime_env.py new file mode 100644 index 00000000..93606d73 --- /dev/null +++ b/src/codeflare_sdk/ray/rayjobs/runtime_env.py @@ -0,0 +1,405 @@ +from __future__ import annotations # Postpone evaluation of annotations + +import logging +import os +import re +import yaml +import zipfile +import base64 +import io +from typing import Dict, Any, Optional, List, TYPE_CHECKING +from codeflare_sdk.common.utils.constants import MOUNT_PATH +from kubernetes import client +from ray.runtime_env import RuntimeEnv + +from codeflare_sdk.ray.rayjobs.config import ManagedClusterConfig +from ...common.kubernetes_cluster.auth import get_api_client + +# Use TYPE_CHECKING to avoid circular import at runtime +if TYPE_CHECKING: + from codeflare_sdk.ray.rayjobs.rayjob import RayJob + +logger = logging.getLogger(__name__) + +# Regex pattern for finding Python files in entrypoint commands +# Matches paths like: test.py, ./test.py, dir/test.py, my-dir/test.py +PYTHON_FILE_PATTERN = r"(?:python\s+)?([./\w/-]+\.py)" + +# Path where working_dir will be unzipped on submitter pod +UNZIP_PATH = "/tmp/rayjob-working-dir" + +# Exclude Jupyter notebook and Markdown files from working directory zips +JUPYTER_NOTEBOOK_PATTERN = r"\.ipynb$" +MARKDOWN_FILE_PATTERN = r"\.md$" + + +def _should_exclude_file(file_path: str) -> bool: + """ + Check if file should be excluded from working directory zip. + Currently excludes: + - Jupyter notebook files (.ipynb) + - Markdown files (.md) + + Args: + file_path: Relative file path within the working directory + + Returns: + True if file should be excluded, False otherwise + """ + return bool( + re.search(JUPYTER_NOTEBOOK_PATTERN, file_path, re.IGNORECASE) + or re.search(MARKDOWN_FILE_PATTERN, file_path, re.IGNORECASE) + ) + + +def _normalize_runtime_env( + runtime_env: Optional[RuntimeEnv], +) -> Optional[Dict[str, Any]]: + if runtime_env is None: + return None + return runtime_env.to_dict() + + +def extract_all_local_files(job: RayJob) -> Optional[Dict[str, str]]: + """ + Prepare local files for Secret upload. + + - If runtime_env has local working_dir: zip entire directory into single file + - If single entrypoint file (no working_dir): extract that file + - If remote working_dir URL: return None (pass through to Ray) + + Returns: + Dict with either: + - {"working_dir.zip": } for zipped directories + - {"script.py": } for single files + - None for remote working_dir or no files + """ + # Convert RuntimeEnv to dict for processing + runtime_env_dict = _normalize_runtime_env(job.runtime_env) + + # If there's a remote working_dir, don't extract local files + if ( + runtime_env_dict + and "working_dir" in runtime_env_dict + and not os.path.isdir(runtime_env_dict["working_dir"]) + ): + logger.info( + f"Remote working_dir detected: {runtime_env_dict['working_dir']}. " + "Skipping local file extraction - using remote source." + ) + return None + + # If there's a local working_dir, zip it + if ( + runtime_env_dict + and "working_dir" in runtime_env_dict + and os.path.isdir(runtime_env_dict["working_dir"]) + ): + working_dir = runtime_env_dict["working_dir"] + logger.info(f"Zipping local working_dir: {working_dir}") + zip_data = _zip_directory(working_dir) + if zip_data: + # Encode zip as base64 for Secret storage + zip_base64 = base64.b64encode(zip_data).decode("utf-8") + return {"working_dir.zip": zip_base64} + + # If no working_dir, check for single entrypoint file + entrypoint_file = _extract_single_entrypoint_file(job) + if entrypoint_file: + return entrypoint_file + + return None + + +def _zip_directory(directory_path: str) -> Optional[bytes]: + """ + Zip entire directory preserving structure, excluding Jupyter notebook and markdown files. + + Args: + directory_path: Path to directory to zip + + Returns: + Bytes of zip file, or None on error + """ + try: + # Create in-memory zip file + zip_buffer = io.BytesIO() + excluded_count = 0 + + with zipfile.ZipFile(zip_buffer, "w", zipfile.ZIP_DEFLATED) as zipf: + # Walk through directory and add all files + for root, dirs, files in os.walk(directory_path): + for file in files: + file_path = os.path.join(root, file) + # Calculate relative path from directory_path + arcname = os.path.relpath(file_path, directory_path) + + # Check if file should be excluded + if _should_exclude_file(arcname): + excluded_count += 1 + logger.debug(f"Excluded from zip: {arcname}") + continue + + zipf.write(file_path, arcname) + logger.debug(f"Added to zip: {arcname}") + + zip_data = zip_buffer.getvalue() + + # Log summary with exclusion count + log_message = ( + f"Successfully zipped directory: {directory_path} ({len(zip_data)} bytes)" + ) + if excluded_count > 0: + log_message += f" - Excluded {excluded_count} file(s) (.ipynb, .md)" + logger.info(log_message) + + return zip_data + + except (IOError, OSError) as e: + logger.error(f"Failed to zip directory {directory_path}: {e}") + return None + + +def _extract_single_entrypoint_file(job: RayJob) -> Optional[Dict[str, str]]: + """ + Extract single Python file from entrypoint if no working_dir specified. + + Returns a dict with metadata about the file path structure so we can + preserve it when mounting via Secret. + + Args: + job: RayJob instance + + Returns: + Dict with special format: {"__entrypoint_path__": path, "filename": content} + This allows us to preserve directory structure when mounting + """ + if not job.entrypoint: + return None + + # Look for Python file in entrypoint + matches = re.findall(PYTHON_FILE_PATTERN, job.entrypoint) + + for file_path in matches: + # Check if it's a local file + if os.path.isfile(file_path): + try: + with open(file_path, "r") as f: + content = f.read() + + # Use basename as key (Secret keys can't have slashes) + # But store the full path for later use in Secret item.path + filename = os.path.basename(file_path) + relative_path = file_path.lstrip("./") + + logger.info(f"Extracted single entrypoint file: {file_path}") + + # Return special format with metadata + return {"__entrypoint_path__": relative_path, filename: content} + + except (IOError, OSError) as e: + logger.warning(f"Could not read entrypoint file {file_path}: {e}") + + return None + + +def process_runtime_env( + job: RayJob, files: Optional[Dict[str, str]] = None +) -> Optional[str]: + """ + Process runtime_env field to handle env_vars, pip dependencies, and working_dir. + + Returns: + Processed runtime environment as YAML string, or None if no processing needed + """ + # Convert RuntimeEnv to dict for processing + runtime_env_dict = _normalize_runtime_env(job.runtime_env) + + processed_env = {} + + # Handle env_vars + if runtime_env_dict and "env_vars" in runtime_env_dict: + processed_env["env_vars"] = runtime_env_dict["env_vars"] + logger.info( + f"Added {len(runtime_env_dict['env_vars'])} environment variables to runtime_env" + ) + + # Handle pip dependencies + if runtime_env_dict and "pip" in runtime_env_dict: + pip_deps = process_pip_dependencies(job, runtime_env_dict["pip"]) + if pip_deps: + processed_env["pip"] = pip_deps + + # Handle working_dir + if runtime_env_dict and "working_dir" in runtime_env_dict: + working_dir = runtime_env_dict["working_dir"] + if os.path.isdir(working_dir): + # Local working directory - will be zipped and unzipped to UNZIP_PATH by submitter pod + processed_env["working_dir"] = UNZIP_PATH + logger.info( + f"Local working_dir will be zipped, mounted, and unzipped to: {UNZIP_PATH}" + ) + else: + # Remote URI (e.g., GitHub, S3) - pass through as-is + processed_env["working_dir"] = working_dir + logger.info(f"Using remote working_dir: {working_dir}") + + # If no working_dir specified but we have files (single file case) + elif not runtime_env_dict or "working_dir" not in runtime_env_dict: + if files and "working_dir.zip" not in files: + # Single file case - mount at MOUNT_PATH + processed_env["working_dir"] = MOUNT_PATH + logger.info(f"Single file will be mounted at: {MOUNT_PATH}") + + # Convert to YAML string if we have any processed environment + if processed_env: + return yaml.dump(processed_env, default_flow_style=False) + + return None + + +def process_pip_dependencies(job: RayJob, pip_spec) -> Optional[List[str]]: + """ + Process pip dependencies from runtime_env. + + Args: + pip_spec: Can be a list of packages, a string path to requirements.txt, or dict + + Returns: + List of pip dependencies + """ + if isinstance(pip_spec, list): + # Already a list of dependencies + logger.info(f"Using provided pip dependencies: {len(pip_spec)} packages") + return pip_spec + elif isinstance(pip_spec, str): + # Assume it's a path to requirements.txt + return parse_requirements_file(pip_spec) + elif isinstance(pip_spec, dict): + # Handle dict format (e.g., {"packages": [...], "pip_check": False}) + if "packages" in pip_spec: + logger.info( + f"Using pip dependencies from dict: {len(pip_spec['packages'])} packages" + ) + return pip_spec["packages"] + + logger.warning(f"Unsupported pip specification format: {type(pip_spec)}") + return None + + +def parse_requirements_file(requirements_path: str) -> Optional[List[str]]: + """ + Parse a requirements.txt file and return list of dependencies. + + Args: + requirements_path: Path to requirements.txt file + + Returns: + List of pip dependencies + """ + if not os.path.isfile(requirements_path): + logger.warning(f"Requirements file not found: {requirements_path}") + return None + + try: + with open(requirements_path, "r") as f: + lines = f.readlines() + + # Parse requirements, filtering out comments and empty lines + requirements = [] + for line in lines: + line = line.strip() + if line and not line.startswith("#"): + requirements.append(line) + + logger.info(f"Parsed {len(requirements)} dependencies from {requirements_path}") + return requirements + + except (IOError, OSError) as e: + logger.warning(f"Could not read requirements file {requirements_path}: {e}") + return None + + +def create_secret_from_spec( + job: RayJob, secret_spec: Dict[str, Any], rayjob_result: Dict[str, Any] = None +) -> str: + """ + Create Secret from specification via Kubernetes API. + + Args: + secret_spec: Secret specification dictionary + rayjob_result: The result from RayJob creation containing UID + + Returns: + str: Name of the created Secret + """ + + secret_name = secret_spec["metadata"]["name"] + + metadata = client.V1ObjectMeta(**secret_spec["metadata"]) + + # Add owner reference to ensure proper cleanup + # We can trust that rayjob_result contains UID since submit_job() only returns + # complete K8s resources or None, and we already validated result exists + logger.info( + f"Adding owner reference to Secret '{secret_name}' with RayJob UID: {rayjob_result['metadata']['uid']}" + ) + metadata.owner_references = [ + client.V1OwnerReference( + api_version="ray.io/v1", + kind="RayJob", + name=job.name, + uid=rayjob_result["metadata"]["uid"], + controller=True, + block_owner_deletion=True, + ) + ] + + # Convert dict spec to V1Secret + # Use stringData instead of data to avoid double base64 encoding + # Our zip files are already base64-encoded, so stringData will handle the final encoding + secret = client.V1Secret( + metadata=metadata, + type=secret_spec.get("type", "Opaque"), + string_data=secret_spec["data"], + ) + + # Create Secret via Kubernetes API + k8s_api = client.CoreV1Api(get_api_client()) + try: + k8s_api.create_namespaced_secret(namespace=job.namespace, body=secret) + logger.info( + f"Created Secret '{secret_name}' with {len(secret_spec['data'])} files" + ) + except client.ApiException as e: + if e.status == 409: # Already exists + logger.info(f"Secret '{secret_name}' already exists, updating...") + k8s_api.replace_namespaced_secret( + name=secret_name, namespace=job.namespace, body=secret + ) + else: + raise RuntimeError(f"Failed to create Secret '{secret_name}': {e}") + + return secret_name + + +def create_file_secret( + job: RayJob, files: Dict[str, str], rayjob_result: Dict[str, Any] +): + """ + Create Secret with owner reference for local files. + """ + # Use a basic config builder for Secret creation + config_builder = ManagedClusterConfig() + + # Filter out metadata keys (like __entrypoint_path__) from Secret data + secret_files = {k: v for k, v in files.items() if not k.startswith("__")} + + # Validate and build Secret spec + config_builder.validate_secret_size(secret_files) + secret_spec = config_builder.build_file_secret_spec( + job_name=job.name, namespace=job.namespace, files=secret_files + ) + + # Create Secret with owner reference + create_secret_from_spec(job, secret_spec, rayjob_result) diff --git a/src/codeflare_sdk/ray/rayjobs/status.py b/src/codeflare_sdk/ray/rayjobs/status.py new file mode 100644 index 00000000..027ed09c --- /dev/null +++ b/src/codeflare_sdk/ray/rayjobs/status.py @@ -0,0 +1,64 @@ +# Copyright 2025 IBM, Red Hat +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +""" +The status sub-module defines Enums containing information for Ray job +deployment states and CodeFlare job states, as well as +dataclasses to store information for Ray jobs. +""" + +from dataclasses import dataclass +from enum import Enum +from typing import Optional + + +class RayJobDeploymentStatus(Enum): + """ + Defines the possible deployment states of a Ray job (from the KubeRay RayJob API). + """ + + COMPLETE = "Complete" + RUNNING = "Running" + FAILED = "Failed" + SUSPENDED = "Suspended" + UNKNOWN = "Unknown" + + +class CodeflareRayJobStatus(Enum): + """ + Defines the possible reportable states of a CodeFlare Ray job. + """ + + COMPLETE = 1 + RUNNING = 2 + FAILED = 3 + SUSPENDED = 4 + UNKNOWN = 5 + + +@dataclass +class RayJobInfo: + """ + For storing information about a Ray job. + """ + + name: str + job_id: str + status: RayJobDeploymentStatus + namespace: str + cluster_name: str + start_time: Optional[str] = None + end_time: Optional[str] = None + failed_attempts: int = 0 + succeeded_attempts: int = 0 diff --git a/src/codeflare_sdk/ray/rayjobs/test/conftest.py b/src/codeflare_sdk/ray/rayjobs/test/conftest.py new file mode 100644 index 00000000..bad195a7 --- /dev/null +++ b/src/codeflare_sdk/ray/rayjobs/test/conftest.py @@ -0,0 +1,45 @@ +"""Shared pytest fixtures for rayjobs tests.""" + +import pytest +from unittest.mock import MagicMock + + +# Global test setup that runs automatically for ALL tests +@pytest.fixture(autouse=True) +def auto_mock_setup(mocker): + """Automatically mock common dependencies for all tests.""" + mocker.patch("kubernetes.config.load_kube_config") + + # Always mock get_default_kueue_name to prevent K8s API calls + mocker.patch( + "codeflare_sdk.ray.rayjobs.rayjob.get_default_kueue_name", + return_value="default-queue", + ) + + mock_get_ns = mocker.patch( + "codeflare_sdk.ray.rayjobs.rayjob.get_current_namespace", + return_value="test-namespace", + ) + + mock_rayjob_api = mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayjobApi") + mock_rayjob_instance = MagicMock() + mock_rayjob_api.return_value = mock_rayjob_instance + + mock_cluster_api = mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayClusterApi") + mock_cluster_instance = MagicMock() + mock_cluster_api.return_value = mock_cluster_instance + + mock_k8s_api = mocker.patch("kubernetes.client.CoreV1Api") + mock_k8s_instance = MagicMock() + mock_k8s_api.return_value = mock_k8s_instance + + # Mock get_api_client in runtime_env module where it's actually used + mocker.patch("codeflare_sdk.ray.rayjobs.runtime_env.get_api_client") + + # Return the mocked instances so tests can configure them as needed + return { + "rayjob_api": mock_rayjob_instance, + "cluster_api": mock_cluster_instance, + "k8s_api": mock_k8s_instance, + "get_current_namespace": mock_get_ns, + } diff --git a/src/codeflare_sdk/ray/rayjobs/test/test_config.py b/src/codeflare_sdk/ray/rayjobs/test/test_config.py new file mode 100644 index 00000000..182ff90c --- /dev/null +++ b/src/codeflare_sdk/ray/rayjobs/test/test_config.py @@ -0,0 +1,346 @@ +""" +Tests for the simplified ManagedClusterConfig accelerator_configs behavior. +""" + +import pytest +from codeflare_sdk.ray.rayjobs.config import ManagedClusterConfig, DEFAULT_ACCELERATORS +from kubernetes.client import V1VolumeMount +from kubernetes.client import V1Volume, V1SecretVolumeSource + + +def test_accelerator_configs_defaults_to_default_accelerators(): + """Test that accelerator_configs defaults to DEFAULT_ACCELERATORS.copy()""" + config = ManagedClusterConfig() + + # Should have all the default accelerators + assert "nvidia.com/gpu" in config.accelerator_configs + assert "intel.com/gpu" in config.accelerator_configs + assert "google.com/tpu" in config.accelerator_configs + + # Should be a copy, not the same object + assert config.accelerator_configs is not DEFAULT_ACCELERATORS + assert config.accelerator_configs == DEFAULT_ACCELERATORS + + +def test_accelerator_configs_can_be_overridden(): + """Test that users can override accelerator_configs with custom mappings""" + custom_configs = { + "nvidia.com/gpu": "GPU", + "custom.com/accelerator": "CUSTOM_ACCELERATOR", + } + + config = ManagedClusterConfig(accelerator_configs=custom_configs) + + # Should have custom configs + assert config.accelerator_configs == custom_configs + assert "custom.com/accelerator" in config.accelerator_configs + assert "nvidia.com/gpu" in config.accelerator_configs + + # Should NOT have other defaults + assert "intel.com/gpu" not in config.accelerator_configs + assert "google.com/tpu" not in config.accelerator_configs + + +def test_accelerator_configs_can_extend_defaults(): + """Test that users can extend defaults by providing additional configs""" + extended_configs = { + **DEFAULT_ACCELERATORS, + "custom.com/accelerator": "CUSTOM_ACCEL", + } + + config = ManagedClusterConfig(accelerator_configs=extended_configs) + + # Should have all defaults plus custom + assert "nvidia.com/gpu" in config.accelerator_configs + assert "intel.com/gpu" in config.accelerator_configs + assert "custom.com/accelerator" in config.accelerator_configs + assert config.accelerator_configs["custom.com/accelerator"] == "CUSTOM_ACCEL" + + +def test_gpu_validation_works_with_defaults(): + """Test that GPU validation works with default accelerator configs""" + config = ManagedClusterConfig(head_accelerators={"nvidia.com/gpu": 1}) + + # Should not raise any errors + assert config.head_accelerators == {"nvidia.com/gpu": 1} + + +def test_gpu_validation_works_with_custom_configs(): + """Test that GPU validation works with custom accelerator configs""" + config = ManagedClusterConfig( + accelerator_configs={"custom.com/accelerator": "CUSTOM_ACCEL"}, + head_accelerators={"custom.com/accelerator": 1}, + ) + + # Should not raise any errors + assert config.head_accelerators == {"custom.com/accelerator": 1} + + +def test_gpu_validation_fails_with_unsupported_accelerator(): + """Test that GPU validation fails with unsupported accelerators""" + with pytest.raises( + ValueError, match="GPU configuration 'unsupported.com/accelerator' not found" + ): + ManagedClusterConfig(head_accelerators={"unsupported.com/accelerator": 1}) + + +def test_config_type_validation_errors(mocker): + """Test that type validation properly raises errors with incorrect types.""" + # Mock the _is_type method to return False for type checking + mocker.patch.object( + ManagedClusterConfig, + "_is_type", + side_effect=lambda value, expected_type: False, # Always fail type check + ) + + # This should raise TypeError during initialization + with pytest.raises(TypeError, match="Type validation failed"): + ManagedClusterConfig() + + +def test_config_is_type_method(): + """Test the _is_type static method for type checking.""" + # Test basic types + assert ManagedClusterConfig._is_type("test", str) is True + assert ManagedClusterConfig._is_type(123, int) is True + assert ManagedClusterConfig._is_type(123, str) is False + + # Test optional types (Union with None) + from typing import Optional + + assert ManagedClusterConfig._is_type(None, Optional[str]) is True + assert ManagedClusterConfig._is_type("test", Optional[str]) is True + assert ManagedClusterConfig._is_type(123, Optional[str]) is False + + # Test dict types + assert ManagedClusterConfig._is_type({}, dict) is True + assert ManagedClusterConfig._is_type({"key": "value"}, dict) is True + assert ManagedClusterConfig._is_type([], dict) is False + + +def test_ray_usage_stats_always_disabled_by_default(): + """Test that RAY_USAGE_STATS_ENABLED is always set to '0' by default""" + config = ManagedClusterConfig() + + # Should always have the environment variable set to "0" + assert "RAY_USAGE_STATS_ENABLED" in config.envs + assert config.envs["RAY_USAGE_STATS_ENABLED"] == "0" + + +def test_ray_usage_stats_overwrites_user_env(): + """Test that RAY_USAGE_STATS_ENABLED is always set to '0' even if user specifies it""" + # User tries to enable usage stats + config = ManagedClusterConfig(envs={"RAY_USAGE_STATS_ENABLED": "1"}) + + # Should still be disabled (our setting takes precedence) + assert "RAY_USAGE_STATS_ENABLED" in config.envs + assert config.envs["RAY_USAGE_STATS_ENABLED"] == "0" + + +def test_ray_usage_stats_overwrites_user_env_string(): + """Test that RAY_USAGE_STATS_ENABLED is always set to '0' even if user specifies it as string""" + # User tries to enable usage stats with string + config = ManagedClusterConfig(envs={"RAY_USAGE_STATS_ENABLED": "true"}) + + # Should still be disabled (our setting takes precedence) + assert "RAY_USAGE_STATS_ENABLED" in config.envs + assert config.envs["RAY_USAGE_STATS_ENABLED"] == "0" + + +def test_ray_usage_stats_with_other_user_envs(): + """Test that RAY_USAGE_STATS_ENABLED is set correctly while preserving other user envs""" + # User sets other environment variables + user_envs = { + "CUSTOM_VAR": "custom_value", + "ANOTHER_VAR": "another_value", + "RAY_USAGE_STATS_ENABLED": "1", # This should be overwritten + } + + config = ManagedClusterConfig(envs=user_envs) + + # Our setting should take precedence + assert config.envs["RAY_USAGE_STATS_ENABLED"] == "0" + + # Other user envs should be preserved + assert config.envs["CUSTOM_VAR"] == "custom_value" + assert config.envs["ANOTHER_VAR"] == "another_value" + + # Total count should be correct (3 user envs) + assert len(config.envs) == 3 + + +def test_add_file_volumes_existing_volume_early_return(): + """Test add_file_volumes early return when volume already exists.""" + + config = ManagedClusterConfig() + + # Pre-add a volume with same name + existing_volume = V1Volume( + name="ray-job-files", + secret=V1SecretVolumeSource(secret_name="existing-files"), + ) + config.volumes.append(existing_volume) + + # Should return early and not add duplicate + config.add_file_volumes(secret_name="new-files") + + # Should still have only one volume, no mount added + assert len(config.volumes) == 1 + assert len(config.volume_mounts) == 0 + + +def test_add_file_volumes_existing_mount_early_return(): + """Test add_file_volumes early return when mount already exists.""" + + config = ManagedClusterConfig() + + # Pre-add a mount with same name + existing_mount = V1VolumeMount(name="ray-job-files", mount_path="/existing/path") + config.volume_mounts.append(existing_mount) + + # Should return early and not add duplicate + config.add_file_volumes(secret_name="new-files") + + # Should still have only one mount, no volume added + assert len(config.volumes) == 0 + assert len(config.volume_mounts) == 1 + + +def test_build_file_secret_spec_labels(): + """Test that build_file_secret_spec creates Secret with correct labels.""" + config = ManagedClusterConfig() + + job_name = "test-job" + namespace = "test-namespace" + files = {"test.py": "print('hello')", "helper.py": "# helper code"} + + secret_spec = config.build_file_secret_spec(job_name, namespace, files) + + assert secret_spec["apiVersion"] == "v1" + assert secret_spec["kind"] == "Secret" + assert secret_spec["type"] == "Opaque" + assert secret_spec["metadata"]["name"] == f"{job_name}-files" + assert secret_spec["metadata"]["namespace"] == namespace + + labels = secret_spec["metadata"]["labels"] + assert labels["ray.io/job-name"] == job_name + assert labels["app.kubernetes.io/managed-by"] == "codeflare-sdk" + assert labels["app.kubernetes.io/component"] == "rayjob-files" + + assert secret_spec["data"] == files + + +def test_managed_cluster_config_uses_update_image_for_head(mocker): + """Test that ManagedClusterConfig calls update_image() for head container.""" + # Mock update_image where it's used (in config module), not where it's defined + mock_update_image = mocker.patch( + "codeflare_sdk.ray.rayjobs.config.update_image", + return_value="mocked-image:latest", + ) + + config = ManagedClusterConfig(image="custom-image:v1") + + # Build cluster spec (which should call update_image) + spec = config.build_ray_cluster_spec("test-cluster") + + # Verify update_image was called for head container + assert mock_update_image.called + # Verify head container has the mocked image + head_container = spec["headGroupSpec"]["template"].spec.containers[0] + assert head_container.image == "mocked-image:latest" + + +def test_managed_cluster_config_uses_update_image_for_worker(mocker): + """Test that ManagedClusterConfig calls update_image() for worker container.""" + # Mock update_image where it's used (in config module), not where it's defined + mock_update_image = mocker.patch( + "codeflare_sdk.ray.rayjobs.config.update_image", + return_value="mocked-image:latest", + ) + + config = ManagedClusterConfig(image="custom-image:v1") + + # Build cluster spec (which should call update_image) + spec = config.build_ray_cluster_spec("test-cluster") + + # Verify update_image was called for worker container + assert mock_update_image.called + # Verify worker container has the mocked image + worker_container = spec["workerGroupSpecs"][0]["template"].spec.containers[0] + assert worker_container.image == "mocked-image:latest" + + +def test_managed_cluster_config_with_empty_image_uses_update_image(mocker): + """Test that empty image triggers update_image() to auto-detect.""" + # Mock update_image where it's used (in config module), not where it's defined + mock_update_image = mocker.patch( + "codeflare_sdk.ray.rayjobs.config.update_image", + return_value="auto-detected-image:py3.12", + ) + + config = ManagedClusterConfig(image="") + + # Build cluster spec + spec = config.build_ray_cluster_spec("test-cluster") + + # Verify update_image was called with empty string + mock_update_image.assert_called_with("") + + # Verify containers have the auto-detected image + head_container = spec["headGroupSpec"]["template"].spec.containers[0] + assert head_container.image == "auto-detected-image:py3.12" + + worker_container = spec["workerGroupSpecs"][0]["template"].spec.containers[0] + assert worker_container.image == "auto-detected-image:py3.12" + + +def test_build_ray_cluster_spec_has_enable_in_tree_autoscaling_false(): + """Test that build_ray_cluster_spec sets enableInTreeAutoscaling to False.""" + config = ManagedClusterConfig() + + spec = config.build_ray_cluster_spec("test-cluster") + + # Verify enableInTreeAutoscaling is set to False (required for Kueue) + assert "enableInTreeAutoscaling" in spec + assert spec["enableInTreeAutoscaling"] is False + + +def test_build_ray_cluster_spec_autoscaling_disabled_for_kueue(): + """Test that autoscaling is explicitly disabled for Kueue-managed jobs.""" + config = ManagedClusterConfig(num_workers=3) + + spec = config.build_ray_cluster_spec("kueue-cluster") + + # Verify enableInTreeAutoscaling is False + assert spec["enableInTreeAutoscaling"] is False + + # Verify worker replicas are fixed (min == max == replicas) + worker_spec = spec["workerGroupSpecs"][0] + assert worker_spec["replicas"] == 3 + assert worker_spec["minReplicas"] == 3 + assert worker_spec["maxReplicas"] == 3 + + +def test_managed_cluster_config_default_image_integration(): + """Test that ManagedClusterConfig works with default images (integration test).""" + # Create config without specifying an image (should auto-detect based on Python version) + config = ManagedClusterConfig() + + # Build cluster spec + spec = config.build_ray_cluster_spec("test-cluster") + + # Verify head container has an image (should be auto-detected) + head_container = spec["headGroupSpec"]["template"].spec.containers[0] + assert head_container.image is not None + assert len(head_container.image) > 0 + # Should be one of the supported images + from codeflare_sdk.common.utils.constants import ( + CUDA_PY311_RUNTIME_IMAGE, + CUDA_PY312_RUNTIME_IMAGE, + ) + + assert head_container.image in [CUDA_PY311_RUNTIME_IMAGE, CUDA_PY312_RUNTIME_IMAGE] + + # Verify worker container has the same image + worker_container = spec["workerGroupSpecs"][0]["template"].spec.containers[0] + assert worker_container.image == head_container.image diff --git a/src/codeflare_sdk/ray/rayjobs/test/test_pretty_print.py b/src/codeflare_sdk/ray/rayjobs/test/test_pretty_print.py new file mode 100644 index 00000000..3bbe8bee --- /dev/null +++ b/src/codeflare_sdk/ray/rayjobs/test/test_pretty_print.py @@ -0,0 +1,265 @@ +# Copyright 2025 IBM, Red Hat +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +from codeflare_sdk.ray.rayjobs.pretty_print import ( + _get_status_display, + print_job_status, + print_no_job_found, +) +from codeflare_sdk.ray.rayjobs.status import RayJobDeploymentStatus, RayJobInfo +from unittest.mock import MagicMock, call + + +def test_get_status_display(): + """ + Test the _get_status_display function. + """ + # Test Complete status + display, color = _get_status_display(RayJobDeploymentStatus.COMPLETE) + assert display == "Complete :white_heavy_check_mark:" + assert color == "[white on green][bold]Name" + + # Test Running status + display, color = _get_status_display(RayJobDeploymentStatus.RUNNING) + assert display == "Running :gear:" + assert color == "[white on blue][bold]Name" + + # Test Failed status + display, color = _get_status_display(RayJobDeploymentStatus.FAILED) + assert display == "Failed :x:" + assert color == "[white on red][bold]Name" + + # Test Suspended status + display, color = _get_status_display(RayJobDeploymentStatus.SUSPENDED) + assert display == "Suspended :pause_button:" + assert color == "[white on yellow][bold]Name" + + # Test Unknown status + display, color = _get_status_display(RayJobDeploymentStatus.UNKNOWN) + assert display == "Unknown :question:" + assert color == "[white on red][bold]Name" + + +def test_print_job_status_running_format(mocker): + """ + Test the print_job_status function format for a running job. + """ + # Mock Rich components to verify format + mock_console = MagicMock() + mock_inner_table = MagicMock() + mock_main_table = MagicMock() + mock_panel = MagicMock() + + # Mock Table to return different instances for inner and main tables + table_instances = [mock_inner_table, mock_main_table] + mock_table_class = MagicMock(side_effect=table_instances) + + mocker.patch( + "codeflare_sdk.ray.rayjobs.pretty_print.Console", return_value=mock_console + ) + mocker.patch("codeflare_sdk.ray.rayjobs.pretty_print.Table", mock_table_class) + mocker.patch("codeflare_sdk.ray.rayjobs.pretty_print.Panel", mock_panel) + + # Create test job info for running job + job_info = RayJobInfo( + name="test-job", + job_id="test-job-abc123", + status=RayJobDeploymentStatus.RUNNING, + namespace="test-ns", + cluster_name="test-cluster", + start_time="2025-07-28T11:37:07Z", + failed_attempts=0, + succeeded_attempts=0, + ) + + # Call the function + print_job_status(job_info) + + # Verify both Table calls + expected_table_calls = [ + call(box=None, show_header=False), # Inner content table + call( + box=None, title="[bold] :package: CodeFlare RayJob Status :package:" + ), # Main wrapper table + ] + mock_table_class.assert_has_calls(expected_table_calls) + + # Verify inner table rows are added in correct order and format (versus our hard-coded version of this for cluster) + expected_calls = [ + call("[white on blue][bold]Name"), # Header with blue color for running + call( + "[bold underline]test-job", "Running :gear:" + ), # Name and status with gear emoji + call(), # Empty separator row + call("[bold]Job ID:[/bold] test-job-abc123"), + call("[bold]Status:[/bold] Running"), + call("[bold]RayCluster:[/bold] test-cluster"), + call("[bold]Namespace:[/bold] test-ns"), + call(), # Empty row before timing info + call("[bold]Started:[/bold] 2025-07-28T11:37:07Z"), + ] + mock_inner_table.add_row.assert_has_calls(expected_calls) + + # Verify Panel is created with inner table + mock_panel.fit.assert_called_once_with(mock_inner_table) + + # Verify main table gets the panel + mock_main_table.add_row.assert_called_once_with(mock_panel.fit.return_value) + + # Verify console prints the main table + mock_console.print.assert_called_once_with(mock_main_table) + + +def test_print_job_status_complete_format(mocker): + """ + Test the print_job_status function format for a completed job. + """ + # Mock Rich components + mock_console = MagicMock() + mock_inner_table = MagicMock() + mock_main_table = MagicMock() + mock_panel = MagicMock() + + # Mock Table to return different instances + table_instances = [mock_inner_table, mock_main_table] + mock_table_class = MagicMock(side_effect=table_instances) + + mocker.patch( + "codeflare_sdk.ray.rayjobs.pretty_print.Console", return_value=mock_console + ) + mocker.patch("codeflare_sdk.ray.rayjobs.pretty_print.Table", mock_table_class) + mocker.patch("codeflare_sdk.ray.rayjobs.pretty_print.Panel", mock_panel) + + # Create test job info for completed job + job_info = RayJobInfo( + name="completed-job", + job_id="completed-job-xyz789", + status=RayJobDeploymentStatus.COMPLETE, + namespace="prod-ns", + cluster_name="prod-cluster", + start_time="2025-07-28T11:37:07Z", + failed_attempts=0, + succeeded_attempts=1, + ) + + # Call the function + print_job_status(job_info) + + # Verify correct header color for completed job (green) (versus our hard-coded version of this for cluster) + expected_calls = [ + call("[white on green][bold]Name"), # Green header for complete + call( + "[bold underline]completed-job", "Complete :white_heavy_check_mark:" + ), # Checkmark emoji + call(), # Empty separator + call("[bold]Job ID:[/bold] completed-job-xyz789"), + call("[bold]Status:[/bold] Complete"), + call("[bold]RayCluster:[/bold] prod-cluster"), + call("[bold]Namespace:[/bold] prod-ns"), + call(), # Empty row before timing info + call("[bold]Started:[/bold] 2025-07-28T11:37:07Z"), + ] + mock_inner_table.add_row.assert_has_calls(expected_calls) + + +def test_print_job_status_failed_with_attempts_format(mocker): + """ + Test the print_job_status function format for a failed job with attempts. + """ + # Mock Rich components + mock_console = MagicMock() + mock_inner_table = MagicMock() + mock_main_table = MagicMock() + mock_panel = MagicMock() + + # Mock Table to return different instances + table_instances = [mock_inner_table, mock_main_table] + mock_table_class = MagicMock(side_effect=table_instances) + + mocker.patch( + "codeflare_sdk.ray.rayjobs.pretty_print.Console", return_value=mock_console + ) + mocker.patch("codeflare_sdk.ray.rayjobs.pretty_print.Table", mock_table_class) + mocker.patch("codeflare_sdk.ray.rayjobs.pretty_print.Panel", mock_panel) + + # Create test job info with failures + job_info = RayJobInfo( + name="failing-job", + job_id="failing-job-fail123", + status=RayJobDeploymentStatus.FAILED, + namespace="test-ns", + cluster_name="test-cluster", + start_time="2025-07-28T11:37:07Z", + failed_attempts=3, # Has failures + succeeded_attempts=0, + ) + + # Call the function + print_job_status(job_info) + + # Verify correct formatting including failure attempts (versus our hard-coded version of this for cluster) + expected_calls = [ + call("[white on red][bold]Name"), # Red header for failed + call("[bold underline]failing-job", "Failed :x:"), # X emoji for failed + call(), # Empty separator + call("[bold]Job ID:[/bold] failing-job-fail123"), + call("[bold]Status:[/bold] Failed"), + call("[bold]RayCluster:[/bold] test-cluster"), + call("[bold]Namespace:[/bold] test-ns"), + call(), # Empty row before timing info + call("[bold]Started:[/bold] 2025-07-28T11:37:07Z"), + call("[bold]Failed Attempts:[/bold] 3"), # Failed attempts should be shown + ] + mock_inner_table.add_row.assert_has_calls(expected_calls) + + +def test_print_no_job_found_format(mocker): + """ + Test the print_no_job_found function format. + """ + # Mock Rich components + mock_console = MagicMock() + mock_inner_table = MagicMock() + mock_main_table = MagicMock() + mock_panel = MagicMock() + + # Mock Table to return different instances + table_instances = [mock_inner_table, mock_main_table] + mock_table_class = MagicMock(side_effect=table_instances) + + mocker.patch( + "codeflare_sdk.ray.rayjobs.pretty_print.Console", return_value=mock_console + ) + mocker.patch("codeflare_sdk.ray.rayjobs.pretty_print.Table", mock_table_class) + mocker.patch("codeflare_sdk.ray.rayjobs.pretty_print.Panel", mock_panel) + + # Call the function + print_no_job_found("missing-job", "test-namespace") + + # Verify error message format (versus our hard-coded version of this for cluster) + expected_calls = [ + call("[white on red][bold]Name"), # Red header for error + call( + "[bold underline]missing-job", "[bold red]No RayJob found" + ), # Error message in red + call(), # Empty separator + call(), # Another empty row + call("Please run rayjob.submit() to submit a job."), # Helpful hint + call(), # Empty separator + call("[bold]Namespace:[/bold] test-namespace"), + ] + mock_inner_table.add_row.assert_has_calls(expected_calls) + + # Verify Panel is used + mock_panel.fit.assert_called_once_with(mock_inner_table) diff --git a/src/codeflare_sdk/ray/rayjobs/test/test_rayjob.py b/src/codeflare_sdk/ray/rayjobs/test/test_rayjob.py new file mode 100644 index 00000000..928cc1f8 --- /dev/null +++ b/src/codeflare_sdk/ray/rayjobs/test/test_rayjob.py @@ -0,0 +1,1783 @@ +# Copyright 2022-2025 IBM, Red Hat +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +import pytest +from unittest.mock import MagicMock +from codeflare_sdk.common.utils.constants import RAY_VERSION +from ray.runtime_env import RuntimeEnv + +from codeflare_sdk.ray.rayjobs.rayjob import RayJob +from codeflare_sdk.ray.cluster.config import ClusterConfiguration +from codeflare_sdk.ray.rayjobs.config import ManagedClusterConfig +from kubernetes.client import V1Volume, V1VolumeMount, V1Toleration + + +def test_rayjob_submit_success(auto_mock_setup): + """ + Test successful RayJob submission. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + + mock_api_instance.submit.return_value = {"metadata": {"name": "test-rayjob"}} + + rayjob = RayJob( + job_name="test-rayjob", + cluster_name="test-ray-cluster", + namespace="test-namespace", + entrypoint="python -c 'print(\"hello world\")'", + runtime_env=RuntimeEnv(pip=["requests"]), + ) + + job_id = rayjob.submit() + + assert job_id == "test-rayjob" + + mock_api_instance.submit_job.assert_called_once() + call_args = mock_api_instance.submit_job.call_args + + assert call_args.kwargs["k8s_namespace"] == "test-namespace" + + job_cr = call_args.kwargs["job"] + assert job_cr["metadata"]["name"] == "test-rayjob" + assert job_cr["metadata"]["namespace"] == "test-namespace" + assert job_cr["spec"]["entrypoint"] == "python -c 'print(\"hello world\")'" + assert job_cr["spec"]["clusterSelector"]["ray.io/cluster"] == "test-ray-cluster" + assert job_cr["spec"]["runtimeEnvYAML"] == "pip:\n- requests\n" + + +def test_rayjob_submit_failure(auto_mock_setup): + """ + Test RayJob submission failure. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + + mock_api_instance.submit_job.return_value = None + + rayjob = RayJob( + job_name="test-rayjob", + cluster_name="test-ray-cluster", + namespace="default", + entrypoint="python -c 'print()'", + runtime_env=RuntimeEnv(pip=["numpy"]), + ) + + with pytest.raises(RuntimeError, match="Failed to submit RayJob test-rayjob"): + rayjob.submit() + + +def test_rayjob_init_validation_both_provided(auto_mock_setup): + """ + Test that providing both cluster_name and cluster_config raises error. + """ + cluster_config = ClusterConfiguration(name="test-cluster", namespace="test") + + with pytest.raises( + ValueError, + match="❌ Configuration Error: You cannot specify both 'cluster_name' and 'cluster_config'", + ): + RayJob( + job_name="test-job", + cluster_name="existing-cluster", + cluster_config=cluster_config, + entrypoint="python -c 'print()'", + ) + + +def test_rayjob_init_validation_neither_provided(auto_mock_setup): + """ + Test that providing neither cluster_name nor cluster_config raises error. + """ + with pytest.raises( + ValueError, + match="❌ Configuration Error: You must provide either 'cluster_name'", + ): + RayJob(job_name="test-job", entrypoint="python test.py") + + +def test_rayjob_init_with_cluster_config(auto_mock_setup): + """ + Test RayJob initialization with cluster configuration for auto-creation. + """ + cluster_config = ClusterConfiguration( + name="auto-cluster", namespace="test-namespace", num_workers=2 + ) + + rayjob = RayJob( + job_name="test-job", + cluster_config=cluster_config, + entrypoint="python -c 'print()'", + namespace="test-namespace", + ) + + assert rayjob.name == "test-job" + assert rayjob.cluster_name == "test-job-cluster" # Generated from job name + assert rayjob._cluster_config == cluster_config + assert rayjob._cluster_name is None + + +def test_rayjob_cluster_name_generation(auto_mock_setup): + """ + Test that cluster names are generated when config has empty name. + """ + cluster_config = ClusterConfiguration( + name="", # Empty name should trigger generation + namespace="test-namespace", + num_workers=1, + ) + + rayjob = RayJob( + job_name="my-job", + cluster_config=cluster_config, + entrypoint="python -c 'print()'", + namespace="test-namespace", + ) + + assert rayjob.cluster_name == "my-job-cluster" + + +def test_rayjob_cluster_config_namespace_none(auto_mock_setup): + """ + Test that cluster config namespace is set when None. + """ + cluster_config = ClusterConfiguration( + name="test-cluster", + namespace=None, # This should be set to job namespace + num_workers=1, + ) + + rayjob = RayJob( + job_name="test-job", + cluster_config=cluster_config, + namespace="job-namespace", + entrypoint="python -c 'print()'", + ) + + assert rayjob.namespace == "job-namespace" + + +def test_rayjob_with_active_deadline_seconds(auto_mock_setup): + """ + Test RayJob CR generation with active deadline seconds. + """ + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + namespace="test-namespace", + entrypoint="python main.py", + active_deadline_seconds=30, + ) + + rayjob_cr = rayjob._build_rayjob_cr() + + assert rayjob_cr["spec"]["activeDeadlineSeconds"] == 30 + + +def test_build_ray_cluster_spec_no_config_error(auto_mock_setup): + """ + Test _build_ray_cluster_spec raises error when no cluster config. + """ + rayjob = RayJob( + job_name="test-job", + cluster_name="existing-cluster", + entrypoint="python -c 'print()'", + namespace="test-namespace", + ) + + rayjob_cr = rayjob._build_rayjob_cr() + + assert rayjob_cr["spec"]["clusterSelector"]["ray.io/cluster"] == "existing-cluster" + assert "rayClusterSpec" not in rayjob_cr["spec"] + + +def test_build_ray_cluster_spec(mocker, auto_mock_setup): + """ + Test _build_ray_cluster_spec method. + """ + + mock_ray_cluster = { + "apiVersion": "ray.io/v1", + "kind": "RayCluster", + "metadata": {"name": "test-cluster", "namespace": "test"}, + "spec": { + "rayVersion": RAY_VERSION, + "headGroupSpec": {"replicas": 1}, + "workerGroupSpecs": [{"replicas": 2}], + }, + } + cluster_config = ManagedClusterConfig(num_workers=2) + mocker.patch.object( + cluster_config, "build_ray_cluster_spec", return_value=mock_ray_cluster["spec"] + ) + + rayjob = RayJob( + job_name="test-job", + cluster_config=cluster_config, + entrypoint="python -c 'print()'", + namespace="test-namespace", + ) + + rayjob_cr = rayjob._build_rayjob_cr() + + assert "rayClusterSpec" in rayjob_cr["spec"] + cluster_config.build_ray_cluster_spec.assert_called_once_with( + cluster_name="test-job-cluster" + ) + + +def test_build_rayjob_cr_with_existing_cluster(auto_mock_setup): + """ + Test _build_rayjob_cr method with existing cluster. + """ + + rayjob = RayJob( + job_name="test-job", + cluster_name="existing-cluster", + namespace="test-namespace", + entrypoint="python main.py", + ttl_seconds_after_finished=300, + ) + + rayjob_cr = rayjob._build_rayjob_cr() + + assert rayjob_cr["apiVersion"] == "ray.io/v1" + assert rayjob_cr["kind"] == "RayJob" + assert rayjob_cr["metadata"]["name"] == "test-job" + spec = rayjob_cr["spec"] + assert spec["entrypoint"] == "python main.py" + assert spec["shutdownAfterJobFinishes"] is False + assert spec["ttlSecondsAfterFinished"] == 300 + + assert spec["clusterSelector"]["ray.io/cluster"] == "existing-cluster" + assert "rayClusterSpec" not in spec + + +def test_build_rayjob_cr_with_auto_cluster(mocker, auto_mock_setup): + """ + Test _build_rayjob_cr method with auto-created cluster. + """ + mock_ray_cluster = { + "apiVersion": "ray.io/v1", + "kind": "RayCluster", + "metadata": {"name": "auto-cluster", "namespace": "test"}, + "spec": { + "rayVersion": RAY_VERSION, + "headGroupSpec": {"replicas": 1}, + "workerGroupSpecs": [{"replicas": 2}], + }, + } + cluster_config = ManagedClusterConfig(num_workers=2) + + mocker.patch.object( + cluster_config, "build_ray_cluster_spec", return_value=mock_ray_cluster["spec"] + ) + + rayjob = RayJob( + job_name="test-job", + cluster_config=cluster_config, + entrypoint="python main.py", + namespace="test-namespace", + ) + + rayjob_cr = rayjob._build_rayjob_cr() + assert rayjob_cr["spec"]["rayClusterSpec"] == mock_ray_cluster["spec"] + assert "clusterSelector" not in rayjob_cr["spec"] + + +def test_submit_validation_no_entrypoint(auto_mock_setup): + """ + Test that submit() raises error when entrypoint is None. + """ + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint=None, # No entrypoint provided + namespace="test-namespace", + ) + + with pytest.raises( + ValueError, match="Entrypoint must be provided to submit a RayJob" + ): + rayjob.submit() + + +def test_submit_with_auto_cluster(mocker, auto_mock_setup): + """ + Test successful submission with auto-created cluster. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + + mock_ray_cluster = { + "apiVersion": "ray.io/v1", + "kind": "RayCluster", + "spec": { + "rayVersion": RAY_VERSION, + "headGroupSpec": {"replicas": 1}, + "workerGroupSpecs": [{"replicas": 1}], + }, + } + mock_api_instance.submit_job.return_value = True + + cluster_config = ManagedClusterConfig(num_workers=1) + mocker.patch.object( + cluster_config, "build_ray_cluster_spec", return_value=mock_ray_cluster["spec"] + ) + + rayjob = RayJob( + job_name="test-job", + cluster_config=cluster_config, + entrypoint="python -c 'print()'", + namespace="test-namespace", + ) + + result = rayjob.submit() + + assert result == "test-job" + + mock_api_instance.submit_job.assert_called_once() + call_args = mock_api_instance.submit_job.call_args + + job_cr = call_args.kwargs["job"] + assert "rayClusterSpec" in job_cr["spec"] + assert job_cr["spec"]["rayClusterSpec"] == mock_ray_cluster["spec"] + + +def test_namespace_auto_detection_success(auto_mock_setup): + """ + Test successful namespace auto-detection. + """ + auto_mock_setup["get_current_namespace"].return_value = "detected-ns" + + rayjob = RayJob( + job_name="test-job", entrypoint="python test.py", cluster_name="test-cluster" + ) + + assert rayjob.namespace == "detected-ns" + + +def test_namespace_auto_detection_fallback(auto_mock_setup): + """ + Test that namespace auto-detection failure raises an error. + """ + auto_mock_setup["get_current_namespace"].return_value = None + + with pytest.raises(ValueError, match="Could not auto-detect Kubernetes namespace"): + RayJob( + job_name="test-job", + entrypoint="python -c 'print()'", + cluster_name="test-cluster", + ) + + +def test_namespace_explicit_override(auto_mock_setup): + """ + Test that explicit namespace overrides auto-detection. + """ + auto_mock_setup["get_current_namespace"].return_value = "detected-ns" + + rayjob = RayJob( + job_name="test-job", + entrypoint="python -c 'print()'", + cluster_name="test-cluster", + namespace="explicit-ns", + ) + + assert rayjob.namespace == "explicit-ns" + + +def test_rayjob_with_rayjob_cluster_config(auto_mock_setup): + """ + Test RayJob with the new ManagedClusterConfig. + """ + cluster_config = ManagedClusterConfig( + num_workers=2, + head_cpu_requests="500m", + head_memory_requests="512Mi", + ) + + rayjob = RayJob( + job_name="test-job", + entrypoint="python -c 'print()'", + cluster_config=cluster_config, + namespace="test-namespace", + ) + + assert rayjob._cluster_config == cluster_config + assert rayjob.cluster_name == "test-job-cluster" # Generated from job name + + +def test_rayjob_cluster_config_validation(auto_mock_setup): + """ + Test validation of ManagedClusterConfig parameters. + """ + cluster_config = ManagedClusterConfig() + + rayjob = RayJob( + job_name="test-job", + entrypoint="python -c 'print()'", + cluster_config=cluster_config, + namespace="test-namespace", + ) + + assert rayjob._cluster_config is not None + + +def test_rayjob_missing_entrypoint_validation(auto_mock_setup): + """ + Test that RayJob requires entrypoint for submission. + """ + with pytest.raises( + TypeError, match="missing 1 required positional argument: 'entrypoint'" + ): + RayJob( + job_name="test-job", + cluster_name="test-cluster", + ) + + +def test_build_ray_cluster_spec_integration(mocker, auto_mock_setup): + """ + Test integration with the new build_ray_cluster_spec method. + """ + cluster_config = ManagedClusterConfig() + mock_spec = {"spec": "test-spec"} + mocker.patch.object( + cluster_config, "build_ray_cluster_spec", return_value=mock_spec + ) + + rayjob = RayJob( + job_name="test-job", + entrypoint="python -c 'print()'", + cluster_config=cluster_config, + namespace="test-namespace", + ) + + rayjob_cr = rayjob._build_rayjob_cr() + + cluster_config.build_ray_cluster_spec.assert_called_once_with( + cluster_name="test-job-cluster" + ) + assert "rayClusterSpec" in rayjob_cr["spec"] + assert rayjob_cr["spec"]["rayClusterSpec"] == mock_spec + + +def test_rayjob_with_runtime_env(auto_mock_setup): + """ + Test RayJob with runtime environment configuration. + """ + runtime_env = RuntimeEnv(pip=["numpy", "pandas"]) + + rayjob = RayJob( + job_name="test-job", + entrypoint="python -c 'print()'", + cluster_name="test-cluster", + runtime_env=runtime_env, + namespace="test-namespace", + ) + + assert rayjob.runtime_env == runtime_env + + rayjob_cr = rayjob._build_rayjob_cr() + assert rayjob_cr["spec"]["runtimeEnvYAML"] == "pip:\n- numpy\n- pandas\n" + + +def test_rayjob_with_runtime_env_dict(auto_mock_setup): + """ + Test RayJob with runtime environment as dict (user convenience). + Users can pass a dict instead of having to import RuntimeEnv. + """ + # User can pass dict instead of RuntimeEnv object + runtime_env_dict = { + "pip": ["numpy", "pandas"], + "env_vars": {"TEST_VAR": "test_value"}, + } + + rayjob = RayJob( + job_name="test-job", + entrypoint="python -c 'print()'", + cluster_name="test-cluster", + runtime_env=runtime_env_dict, + namespace="test-namespace", + ) + + # Should be converted to RuntimeEnv internally + assert isinstance(rayjob.runtime_env, RuntimeEnv) + assert rayjob.runtime_env["env_vars"] == {"TEST_VAR": "test_value"} + + # Verify it generates proper YAML output + rayjob_cr = rayjob._build_rayjob_cr() + assert "runtimeEnvYAML" in rayjob_cr["spec"] + runtime_yaml = rayjob_cr["spec"]["runtimeEnvYAML"] + assert "pip:" in runtime_yaml or "pip_packages:" in runtime_yaml + assert "env_vars:" in runtime_yaml + assert "TEST_VAR" in runtime_yaml + + +def test_rayjob_with_active_deadline_and_ttl(auto_mock_setup): + """ + Test RayJob with both active deadline and TTL settings. + """ + + rayjob = RayJob( + job_name="test-job", + entrypoint="python -c 'print()'", + cluster_name="test-cluster", + active_deadline_seconds=300, + ttl_seconds_after_finished=600, + namespace="test-namespace", + ) + + assert rayjob.active_deadline_seconds == 300 + assert rayjob.ttl_seconds_after_finished == 600 + + rayjob_cr = rayjob._build_rayjob_cr() + assert rayjob_cr["spec"]["activeDeadlineSeconds"] == 300 + assert rayjob_cr["spec"]["ttlSecondsAfterFinished"] == 600 + + +def test_rayjob_cluster_name_generation_with_config(auto_mock_setup): + """ + Test cluster name generation when using cluster_config. + """ + + cluster_config = ManagedClusterConfig() + + rayjob = RayJob( + job_name="my-job", + entrypoint="python -c 'print()'", + cluster_config=cluster_config, + namespace="test-namespace", # Explicitly specify namespace + ) + + assert rayjob.cluster_name == "my-job-cluster" + + +def test_rayjob_namespace_propagation_to_cluster_config(auto_mock_setup): + """ + Test that job namespace is propagated to cluster config when None. + """ + auto_mock_setup["get_current_namespace"].return_value = "detected-ns" + + cluster_config = ManagedClusterConfig() + + rayjob = RayJob( + job_name="test-job", + entrypoint="python -c 'print()'", + cluster_config=cluster_config, + ) + + assert rayjob.namespace == "detected-ns" + + +def test_rayjob_error_handling_invalid_cluster_config(auto_mock_setup): + """ + Test error handling with invalid cluster configuration. + """ + + with pytest.raises(ValueError): + RayJob( + job_name="test-job", + entrypoint="python -c 'print()'", + ) + + +def test_rayjob_constructor_parameter_validation(auto_mock_setup): + """ + Test constructor parameter validation. + """ + rayjob = RayJob( + job_name="test-job", + entrypoint="python -c 'print()'", + cluster_name="test-cluster", + namespace="test-ns", + runtime_env=RuntimeEnv(pip=["numpy"]), + ttl_seconds_after_finished=300, + active_deadline_seconds=600, + ) + + assert rayjob.name == "test-job" + assert rayjob.entrypoint == "python -c 'print()'" + assert rayjob.cluster_name == "test-cluster" + assert rayjob.namespace == "test-ns" + # Check that runtime_env is a RuntimeEnv object and contains pip dependencies + assert isinstance(rayjob.runtime_env, RuntimeEnv) + runtime_env_dict = rayjob.runtime_env.to_dict() + assert "pip" in runtime_env_dict + # Ray transforms pip to dict format with 'packages' key + assert runtime_env_dict["pip"]["packages"] == ["numpy"] + assert rayjob.ttl_seconds_after_finished == 300 + assert rayjob.active_deadline_seconds == 600 + + +def test_build_ray_cluster_spec_function(): + """ + Test the build_ray_cluster_spec method directly. + """ + cluster_config = ManagedClusterConfig( + num_workers=2, + head_cpu_requests="500m", + head_memory_requests="512Mi", + worker_cpu_requests="250m", + worker_memory_requests="256Mi", + ) + + spec = cluster_config.build_ray_cluster_spec("test-cluster") + assert "rayVersion" in spec + assert "enableInTreeAutoscaling" in spec + assert spec["enableInTreeAutoscaling"] is False # Required for Kueue + assert "headGroupSpec" in spec + assert "workerGroupSpecs" in spec + + head_spec = spec["headGroupSpec"] + assert head_spec["serviceType"] == "ClusterIP" + assert head_spec["enableIngress"] is False + assert "rayStartParams" in head_spec + assert "template" in head_spec + worker_specs = spec["workerGroupSpecs"] + assert len(worker_specs) == 1 + worker_spec = worker_specs[0] + assert worker_spec["replicas"] == 2 + assert worker_spec["minReplicas"] == 2 + assert worker_spec["maxReplicas"] == 2 + assert worker_spec["groupName"] == "worker-group-test-cluster" + + +def test_build_ray_cluster_spec_with_accelerators(): + """ + Test build_ray_cluster_spec with GPU accelerators. + """ + cluster_config = ManagedClusterConfig( + head_accelerators={"nvidia.com/gpu": 1}, + worker_accelerators={"nvidia.com/gpu": 2}, + ) + + spec = cluster_config.build_ray_cluster_spec("test-cluster") + head_spec = spec["headGroupSpec"] + head_params = head_spec["rayStartParams"] + assert "num-gpus" in head_params + assert head_params["num-gpus"] == "1" + + worker_specs = spec["workerGroupSpecs"] + worker_spec = worker_specs[0] + worker_params = worker_spec["rayStartParams"] + assert "num-gpus" in worker_params + assert worker_params["num-gpus"] == "2" + + +def test_build_ray_cluster_spec_with_custom_volumes(): + """ + Test build_ray_cluster_spec with custom volumes and volume mounts. + """ + custom_volume = V1Volume(name="custom-data", empty_dir={}) + custom_volume_mount = V1VolumeMount(name="custom-data", mount_path="/data") + cluster_config = ManagedClusterConfig( + volumes=[custom_volume], + volume_mounts=[custom_volume_mount], + ) + + spec = cluster_config.build_ray_cluster_spec("test-cluster") + head_spec = spec["headGroupSpec"] + head_pod_spec = head_spec["template"].spec + assert len(head_pod_spec.volumes) > 0 + + head_container = head_pod_spec.containers[0] + assert len(head_container.volume_mounts) > 0 + + +def test_build_ray_cluster_spec_with_environment_variables(): + """ + Test build_ray_cluster_spec with environment variables. + """ + cluster_config = ManagedClusterConfig( + envs={"CUDA_VISIBLE_DEVICES": "0", "RAY_DISABLE_IMPORT_WARNING": "1"}, + ) + + spec = cluster_config.build_ray_cluster_spec("test-cluster") + + head_spec = spec["headGroupSpec"] + head_pod_spec = head_spec["template"].spec + head_container = head_pod_spec.containers[0] + assert hasattr(head_container, "env") + env_vars = {env.name: env.value for env in head_container.env} + assert env_vars["CUDA_VISIBLE_DEVICES"] == "0" + assert env_vars["RAY_DISABLE_IMPORT_WARNING"] == "1" + worker_specs = spec["workerGroupSpecs"] + worker_spec = worker_specs[0] + worker_pod_spec = worker_spec["template"].spec + worker_container = worker_pod_spec.containers[0] + + assert hasattr(worker_container, "env") + worker_env_vars = {env.name: env.value for env in worker_container.env} + assert worker_env_vars["CUDA_VISIBLE_DEVICES"] == "0" + assert worker_env_vars["RAY_DISABLE_IMPORT_WARNING"] == "1" + + +def test_build_ray_cluster_spec_with_tolerations(): + """ + Test build_ray_cluster_spec with tolerations. + """ + head_toleration = V1Toleration( + key="node-role.kubernetes.io/master", operator="Exists", effect="NoSchedule" + ) + worker_toleration = V1Toleration( + key="nvidia.com/gpu", operator="Exists", effect="NoSchedule" + ) + + cluster_config = ManagedClusterConfig( + head_tolerations=[head_toleration], + worker_tolerations=[worker_toleration], + ) + + spec = cluster_config.build_ray_cluster_spec("test-cluster") + head_spec = spec["headGroupSpec"] + head_pod_spec = head_spec["template"].spec + assert hasattr(head_pod_spec, "tolerations") + assert len(head_pod_spec.tolerations) == 1 + assert head_pod_spec.tolerations[0].key == "node-role.kubernetes.io/master" + + worker_specs = spec["workerGroupSpecs"] + worker_spec = worker_specs[0] + worker_pod_spec = worker_spec["template"].spec + assert hasattr(worker_pod_spec, "tolerations") + assert len(worker_pod_spec.tolerations) == 1 + assert worker_pod_spec.tolerations[0].key == "nvidia.com/gpu" + + +def test_build_ray_cluster_spec_with_image_pull_secrets(): + """ + Test build_ray_cluster_spec with image pull secrets. + """ + cluster_config = ManagedClusterConfig( + image_pull_secrets=["my-registry-secret", "another-secret"] + ) + + spec = cluster_config.build_ray_cluster_spec("test-cluster") + + head_spec = spec["headGroupSpec"] + head_pod_spec = head_spec["template"].spec + assert hasattr(head_pod_spec, "image_pull_secrets") + + head_secrets = head_pod_spec.image_pull_secrets + assert len(head_secrets) == 2 + assert head_secrets[0].name == "my-registry-secret" + assert head_secrets[1].name == "another-secret" + + worker_specs = spec["workerGroupSpecs"] + worker_spec = worker_specs[0] + worker_pod_spec = worker_spec["template"].spec + assert hasattr(worker_pod_spec, "image_pull_secrets") + + worker_secrets = worker_pod_spec.image_pull_secrets + assert len(worker_secrets) == 2 + assert worker_secrets[0].name == "my-registry-secret" + assert worker_secrets[1].name == "another-secret" + + +def test_submit_with_cluster_config_compatible_image_passes(auto_mock_setup): + """ + Test that submission passes with compatible cluster_config image. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + mock_api_instance.submit_job.return_value = True + + cluster_config = ManagedClusterConfig(image=f"ray:{RAY_VERSION}") + + rayjob = RayJob( + job_name="test-job", + cluster_config=cluster_config, + namespace="test-namespace", + entrypoint="python -c 'print()'", + ) + + result = rayjob.submit() + assert result == "test-job" + + +def test_submit_with_cluster_config_incompatible_image_fails(auto_mock_setup): + """ + Test that submission fails with incompatible cluster_config image. + """ + + cluster_config = ManagedClusterConfig(image="ray:2.8.0") # Different version + + rayjob = RayJob( + job_name="test-job", + cluster_config=cluster_config, + namespace="test-namespace", + entrypoint="python -c 'print()'", + ) + + with pytest.raises( + ValueError, match="Cluster config image: Ray version mismatch detected" + ): + rayjob.submit() + + +def test_validate_ray_version_compatibility_method(auto_mock_setup): + """ + Test the _validate_ray_version_compatibility method directly. + """ + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + namespace="test-namespace", + entrypoint="python -c 'print()'", + ) + + rayjob._validate_ray_version_compatibility() + rayjob._cluster_config = ManagedClusterConfig(image=f"ray:{RAY_VERSION}") + rayjob._validate_ray_version_compatibility() + rayjob._cluster_config = ManagedClusterConfig(image="ray:2.8.0") + with pytest.raises( + ValueError, match="Cluster config image: Ray version mismatch detected" + ): + rayjob._validate_ray_version_compatibility() + + rayjob._cluster_config = ManagedClusterConfig(image="custom-image:latest") + with pytest.warns( + UserWarning, match="Cluster config image: Cannot determine Ray version" + ): + rayjob._validate_ray_version_compatibility() + + +def test_validate_cluster_config_image_method(auto_mock_setup): + """ + Test the _validate_cluster_config_image method directly. + """ + + rayjob = RayJob( + job_name="test-job", + cluster_config=ManagedClusterConfig(), + namespace="test-namespace", + entrypoint="python -c 'print()'", + ) + + rayjob._validate_cluster_config_image() + rayjob._cluster_config.image = f"ray:{RAY_VERSION}" + rayjob._validate_cluster_config_image() + rayjob._cluster_config.image = "ray:2.8.0" + with pytest.raises( + ValueError, match="Cluster config image: Ray version mismatch detected" + ): + rayjob._validate_cluster_config_image() + + rayjob._cluster_config.image = "custom-image:latest" + with pytest.warns( + UserWarning, match="Cluster config image: Cannot determine Ray version" + ): + rayjob._validate_cluster_config_image() + + +def test_validate_cluster_config_image_edge_cases(auto_mock_setup): + """ + Test edge cases in _validate_cluster_config_image method. + """ + + rayjob = RayJob( + job_name="test-job", + cluster_config=ManagedClusterConfig(), + namespace="test-namespace", + entrypoint="python -c 'print()'", + ) + + rayjob._cluster_config.image = None + rayjob._validate_cluster_config_image() + rayjob._cluster_config.image = "" + rayjob._validate_cluster_config_image() + rayjob._cluster_config.image = 123 + rayjob._validate_cluster_config_image() + + class MockClusterConfig: + pass + + rayjob._cluster_config = MockClusterConfig() + rayjob._validate_cluster_config_image() + + +def test_rayjob_stop_success(auto_mock_setup, caplog): + """ + Test successful RayJob stop operation. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + + mock_api_instance.suspend_job.return_value = { + "metadata": {"name": "test-rayjob"}, + "spec": {"suspend": True}, + } + + rayjob = RayJob( + job_name="test-rayjob", + cluster_name="test-cluster", + namespace="test-namespace", + entrypoint="python -c 'print()'", + ) + + with caplog.at_level("INFO"): + result = rayjob.stop() + + assert result is True + + mock_api_instance.suspend_job.assert_called_once_with( + name="test-rayjob", k8s_namespace="test-namespace" + ) + + # Verify success message was logged + assert "Successfully stopped the RayJob test-rayjob" in caplog.text + + +def test_rayjob_stop_failure(auto_mock_setup): + """ + Test RayJob stop operation when API call fails. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + + mock_api_instance.suspend_job.return_value = None + + rayjob = RayJob( + job_name="test-rayjob", + cluster_name="test-cluster", + namespace="test-namespace", + entrypoint="python -c 'print()'", + ) + + with pytest.raises(RuntimeError, match="Failed to stop the RayJob test-rayjob"): + rayjob.stop() + + mock_api_instance.suspend_job.assert_called_once_with( + name="test-rayjob", k8s_namespace="test-namespace" + ) + + +def test_rayjob_resubmit_success(auto_mock_setup): + """ + Test successful RayJob resubmit operation. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + + mock_api_instance.resubmit_job.return_value = { + "metadata": {"name": "test-rayjob"}, + "spec": {"suspend": False}, + } + + rayjob = RayJob( + job_name="test-rayjob", + cluster_name="test-cluster", + namespace="test-namespace", + entrypoint="python -c 'print()'", + ) + + result = rayjob.resubmit() + + assert result is True + + mock_api_instance.resubmit_job.assert_called_once_with( + name="test-rayjob", k8s_namespace="test-namespace" + ) + + +def test_rayjob_resubmit_failure(auto_mock_setup): + """ + Test RayJob resubmit operation when API call fails. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + + mock_api_instance.resubmit_job.return_value = None + + rayjob = RayJob( + job_name="test-rayjob", + cluster_name="test-cluster", + namespace="test-namespace", + entrypoint="python -c 'print()'", + ) + + with pytest.raises(RuntimeError, match="Failed to resubmit the RayJob test-rayjob"): + rayjob.resubmit() + + mock_api_instance.resubmit_job.assert_called_once_with( + name="test-rayjob", k8s_namespace="test-namespace" + ) + + +def test_rayjob_delete_success(auto_mock_setup): + """ + Test successful RayJob deletion. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + + rayjob = RayJob( + job_name="test-rayjob", + entrypoint="python -c 'print()'", + cluster_name="test-cluster", + ) + + mock_api_instance.delete_job.return_value = True + + result = rayjob.delete() + + assert result is True + mock_api_instance.delete_job.assert_called_once_with( + name="test-rayjob", k8s_namespace="test-namespace" + ) + + +def test_rayjob_delete_already_deleted(auto_mock_setup, caplog): + """ + Test RayJob deletion when already deleted (should succeed gracefully). + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + + rayjob = RayJob( + job_name="test-rayjob", + entrypoint="python -c 'print()'", + cluster_name="test-cluster", + ) + + # Python client returns False when job doesn't exist/already deleted + mock_api_instance.delete_job.return_value = False + + with caplog.at_level("INFO"): + result = rayjob.delete() + + # Should succeed (not raise error) when already deleted + assert result is True + assert "already deleted or does not exist" in caplog.text + + mock_api_instance.delete_job.assert_called_once_with( + name="test-rayjob", k8s_namespace="test-namespace" + ) + + +def test_rayjob_init_both_none_error(auto_mock_setup): + """ + Test RayJob initialization error when both cluster_name and cluster_config are None. + """ + with pytest.raises( + ValueError, + match="Configuration Error: You must provide either 'cluster_name' .* or 'cluster_config'", + ): + RayJob( + job_name="test-job", + entrypoint="python -c 'print()'", + cluster_name=None, + cluster_config=None, + ) + + +def test_rayjob_init_missing_cluster_name_with_no_config(auto_mock_setup): + """ + Test RayJob initialization error when cluster_name is None without cluster_config. + """ + with pytest.raises( + ValueError, + match="Configuration Error: a 'cluster_name' is required when not providing 'cluster_config'", + ): + rayjob = RayJob.__new__(RayJob) + rayjob.name = "test-job" + rayjob.entrypoint = "python test.py" + rayjob.runtime_env = None + rayjob.ttl_seconds_after_finished = 0 + rayjob.active_deadline_seconds = None + rayjob.shutdown_after_job_finishes = True + rayjob.namespace = "test-namespace" + rayjob._cluster_name = None + rayjob._cluster_config = None + if rayjob._cluster_config is None and rayjob._cluster_name is None: + raise ValueError( + "❌ Configuration Error: a 'cluster_name' is required when not providing 'cluster_config'" + ) + + +def test_rayjob_kueue_label_no_default_queue(auto_mock_setup, mocker, caplog): + """ + Test RayJob falls back to 'default' queue when no default queue exists. + """ + mocker.patch( + "codeflare_sdk.ray.rayjobs.rayjob.get_default_kueue_name", + return_value=None, + ) + + mock_api_instance = auto_mock_setup["rayjob_api"] + mock_api_instance.submit_job.return_value = {"metadata": {"name": "test-job"}} + + cluster_config = ManagedClusterConfig() + rayjob = RayJob( + job_name="test-job", + cluster_config=cluster_config, + entrypoint="python -c 'print()'", + ) + + with caplog.at_level("WARNING"): + rayjob.submit() + + # Verify the submitted job has the fallback label + call_args = mock_api_instance.submit_job.call_args + submitted_job = call_args.kwargs["job"] + assert submitted_job["metadata"]["labels"]["kueue.x-k8s.io/queue-name"] == "default" + + # Verify warning was logged + assert "No default Kueue LocalQueue found" in caplog.text + + +def test_rayjob_kueue_explicit_local_queue(auto_mock_setup): + """ + Test RayJob uses explicitly specified local queue. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + mock_api_instance.submit_job.return_value = {"metadata": {"name": "test-job"}} + + cluster_config = ManagedClusterConfig() + rayjob = RayJob( + job_name="test-job", + cluster_config=cluster_config, + entrypoint="python -c 'print()'", + local_queue="custom-queue", + ) + + rayjob.submit() + + # Verify the submitted job has the explicit queue label + call_args = mock_api_instance.submit_job.call_args + submitted_job = call_args.kwargs["job"] + assert ( + submitted_job["metadata"]["labels"]["kueue.x-k8s.io/queue-name"] + == "custom-queue" + ) + + +def test_rayjob_no_kueue_label_for_existing_cluster(auto_mock_setup): + """ + Test RayJob doesn't add Kueue label for existing clusters. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + mock_api_instance.submit_job.return_value = {"metadata": {"name": "test-job"}} + + # Using existing cluster (no cluster_config) + rayjob = RayJob( + job_name="test-job", + cluster_name="existing-cluster", + entrypoint="python -c 'print()'", + ) + + rayjob.submit() + + # Verify no Kueue label was added + call_args = mock_api_instance.submit_job.call_args + submitted_job = call_args.kwargs["job"] + assert "kueue.x-k8s.io/queue-name" not in submitted_job["metadata"]["labels"] + + +def test_rayjob_with_ttl_and_deadline(auto_mock_setup): + """ + Test RayJob with TTL and active deadline seconds. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + mock_api_instance.submit_job.return_value = {"metadata": {"name": "test-job"}} + + cluster_config = ManagedClusterConfig() + rayjob = RayJob( + job_name="test-job", + cluster_config=cluster_config, + entrypoint="python -c 'print()'", + ttl_seconds_after_finished=300, + active_deadline_seconds=600, + ) + + rayjob.submit() + + # Verify TTL and deadline were set + call_args = mock_api_instance.submit_job.call_args + submitted_job = call_args.kwargs["job"] + assert submitted_job["spec"]["ttlSecondsAfterFinished"] == 300 + assert submitted_job["spec"]["activeDeadlineSeconds"] == 600 + + +def test_rayjob_shutdown_after_job_finishes(auto_mock_setup): + """ + Test RayJob sets shutdownAfterJobFinishes correctly. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + mock_api_instance.submit_job.return_value = {"metadata": {"name": "test-job"}} + + # Test with managed cluster (should shutdown) + cluster_config = ManagedClusterConfig() + rayjob = RayJob( + job_name="test-job", + cluster_config=cluster_config, + entrypoint="python -c 'print()'", + ) + + rayjob.submit() + + call_args = mock_api_instance.submit_job.call_args + submitted_job = call_args.kwargs["job"] + assert submitted_job["spec"]["shutdownAfterJobFinishes"] is True + + # Test with existing cluster (should not shutdown) + rayjob2 = RayJob( + job_name="test-job2", + cluster_name="existing-cluster", + entrypoint="python -c 'print()'", + ) + + rayjob2.submit() + + call_args2 = mock_api_instance.submit_job.call_args + submitted_job2 = call_args2.kwargs["job"] + assert submitted_job2["spec"]["shutdownAfterJobFinishes"] is False + + +def test_rayjob_stop_delete_resubmit_logging(auto_mock_setup, caplog): + """ + Test logging for stop, delete, and resubmit operations. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + + # Test stop with logging + mock_api_instance.suspend_job.return_value = { + "metadata": {"name": "test-rayjob"}, + "spec": {"suspend": True}, + } + + rayjob = RayJob( + job_name="test-rayjob", + cluster_name="test-cluster", + namespace="test-namespace", + entrypoint="python -c 'print()'", + ) + + with caplog.at_level("INFO"): + result = rayjob.stop() + + assert result is True + assert "Successfully stopped the RayJob test-rayjob" in caplog.text + + # Test delete with logging + caplog.clear() + mock_api_instance.delete_job.return_value = True + + with caplog.at_level("INFO"): + result = rayjob.delete() + + assert result is True + assert "Successfully deleted the RayJob test-rayjob" in caplog.text + + # Test resubmit with logging + caplog.clear() + mock_api_instance.resubmit_job.return_value = { + "metadata": {"name": "test-rayjob"}, + "spec": {"suspend": False}, + } + + with caplog.at_level("INFO"): + result = rayjob.resubmit() + + assert result is True + assert "Successfully resubmitted the RayJob test-rayjob" in caplog.text + + +def test_rayjob_initialization_logging(auto_mock_setup, caplog): + """ + Test RayJob initialization logging. + """ + with caplog.at_level("INFO"): + cluster_config = ManagedClusterConfig() + rayjob = RayJob( + job_name="test-job", + cluster_config=cluster_config, + entrypoint="python -c 'print()'", + ) + + assert "Creating new cluster: test-job-cluster" in caplog.text + assert "Initialized RayJob: test-job in namespace: test-namespace" in caplog.text + + +def test_build_submitter_pod_template_uses_default_image(auto_mock_setup, mocker): + """ + Test that _build_submitter_pod_template() uses get_ray_image_for_python_version() for default image. + """ + # Mock get_ray_image_for_python_version to verify it's called + mock_get_image = mocker.patch( + "codeflare_sdk.ray.rayjobs.rayjob.get_ray_image_for_python_version", + return_value="auto-detected-image:py3.12", + ) + + rayjob = RayJob( + job_name="test-job", + cluster_name="existing-cluster", + entrypoint="python -c 'print()'", + namespace="test-namespace", + ) + + files = {"test.py": "print('hello')"} + secret_name = "test-files" + + # Call _build_submitter_pod_template + submitter_template = rayjob._build_submitter_pod_template(files, secret_name) + + # Verify get_ray_image_for_python_version was called + mock_get_image.assert_called_once() + + # Verify the submitter pod uses the auto-detected image + assert ( + submitter_template["spec"]["containers"][0]["image"] + == "auto-detected-image:py3.12" + ) + + +def test_build_submitter_pod_template_uses_cluster_config_image( + auto_mock_setup, mocker +): + """ + Test that _build_submitter_pod_template() uses cluster_config image when provided. + """ + # Mock get_ray_image_for_python_version (should be called but overridden) + mock_get_image = mocker.patch( + "codeflare_sdk.ray.rayjobs.rayjob.get_ray_image_for_python_version", + return_value="auto-detected-image:py3.12", + ) + + cluster_config = ManagedClusterConfig(image="custom-cluster-image:v1") + + rayjob = RayJob( + job_name="test-job", + cluster_config=cluster_config, + entrypoint="python -c 'print()'", + namespace="test-namespace", + ) + + files = {"test.py": "print('hello')"} + secret_name = "test-files" + + # Call _build_submitter_pod_template + submitter_template = rayjob._build_submitter_pod_template(files, secret_name) + + # Verify get_ray_image_for_python_version was called + mock_get_image.assert_called_once() + + # Verify the submitter pod uses the cluster config image (overrides default) + assert ( + submitter_template["spec"]["containers"][0]["image"] + == "custom-cluster-image:v1" + ) + + +def test_build_submitter_pod_template_with_files(auto_mock_setup): + """ + Test that _build_submitter_pod_template() correctly builds Secret items for files. + """ + rayjob = RayJob( + job_name="test-job", + cluster_name="existing-cluster", + entrypoint="python -c 'print()'", + namespace="test-namespace", + ) + + files = {"main.py": "print('main')", "helper.py": "print('helper')"} + secret_name = "test-files" + + # Call _build_submitter_pod_template + submitter_template = rayjob._build_submitter_pod_template(files, secret_name) + + # Verify Secret items are created for each file + secret_items = submitter_template["spec"]["volumes"][0]["secret"]["items"] + assert len(secret_items) == 2 + + # Verify each file has a Secret item + file_names = [item["key"] for item in secret_items] + assert "main.py" in file_names + assert "helper.py" in file_names + + # Verify paths match keys + for item in secret_items: + assert item["key"] == item["path"] + + +def test_validate_working_dir_entrypoint_no_runtime_env(auto_mock_setup, tmp_path): + """ + Test validation checks file exists even when no runtime_env is specified. + """ + # Create the script file + script_file = tmp_path / "script.py" + script_file.write_text("print('hello')") + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint=f"python {script_file}", + namespace="test-namespace", + ) + + # Should not raise exception (file exists) + rayjob._validate_working_dir_entrypoint() + + +def test_validate_working_dir_entrypoint_no_working_dir(auto_mock_setup, tmp_path): + """ + Test validation checks file when runtime_env has no working_dir. + """ + # Create the script file + script_file = tmp_path / "script.py" + script_file.write_text("print('hello')") + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint=f"python {script_file}", + namespace="test-namespace", + runtime_env=RuntimeEnv(pip=["numpy"]), + ) + + # Should not raise exception (file exists) + rayjob._validate_working_dir_entrypoint() + + +def test_validate_working_dir_entrypoint_remote_working_dir(auto_mock_setup): + """ + Test validation skips ALL checks for remote working_dir. + """ + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint="python nonexistent_script.py", # File doesn't exist, but should be ignored + namespace="test-namespace", + runtime_env=RuntimeEnv( + working_dir="https://github.com/user/repo/archive/main.zip" + ), + ) + + # Should not raise any exception (remote working_dir skips all validation) + rayjob._validate_working_dir_entrypoint() + + +def test_validate_working_dir_entrypoint_no_python_file(auto_mock_setup): + """ + Test validation passes when entrypoint has no Python file reference. + """ + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint="echo 'hello world'", # No Python file + namespace="test-namespace", + runtime_env=RuntimeEnv(working_dir="."), + ) + + # Should not raise any exception + rayjob._validate_working_dir_entrypoint() + + +def test_validate_working_dir_entrypoint_no_redundancy(auto_mock_setup, tmp_path): + """ + Test validation passes when entrypoint doesn't reference working_dir. + """ + # Create test directory and file + test_dir = tmp_path / "testdir" + test_dir.mkdir() + script_file = test_dir / "script.py" + script_file.write_text("print('hello')") + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint="python script.py", # No directory prefix + namespace="test-namespace", + runtime_env=RuntimeEnv(working_dir=str(test_dir)), + ) + + # Should not raise any exception + rayjob._validate_working_dir_entrypoint() + + +def test_validate_working_dir_entrypoint_redundant_reference_error( + auto_mock_setup, tmp_path +): + """ + Test validation raises error when entrypoint redundantly references working_dir. + """ + # Create test directory and file + test_dir = tmp_path / "testdir" + test_dir.mkdir() + script_file = test_dir / "script.py" + script_file.write_text("print('hello')") + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint=f"python {test_dir}/script.py", # Redundant reference + namespace="test-namespace", + runtime_env=RuntimeEnv(working_dir=str(test_dir)), + ) + + # Should raise ValueError with helpful message + with pytest.raises(ValueError, match="Working directory conflict detected"): + rayjob._validate_working_dir_entrypoint() + + +def test_validate_working_dir_entrypoint_with_dot_slash(auto_mock_setup, tmp_path): + """ + Test validation handles paths with ./ prefix correctly. + """ + # Create test directory and file + test_dir = tmp_path / "testdir" + test_dir.mkdir() + script_file = test_dir / "script.py" + script_file.write_text("print('hello')") + + # Change to temp directory so relative paths work + import os + + original_cwd = os.getcwd() + try: + os.chdir(tmp_path) + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint="python ./testdir/script.py", # With ./ prefix + namespace="test-namespace", + runtime_env=RuntimeEnv(working_dir="./testdir"), # With ./ prefix + ) + + # Should raise ValueError (redundant reference) + with pytest.raises(ValueError, match="Working directory conflict detected"): + rayjob._validate_working_dir_entrypoint() + finally: + os.chdir(original_cwd) + + +def test_validate_working_dir_entrypoint_submit_integration(auto_mock_setup, tmp_path): + """ + Test that validation is called during submit() and blocks submission. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + mock_api_instance.submit_job.return_value = {"metadata": {"name": "test-job"}} + + # Create test directory and file + test_dir = tmp_path / "testdir" + test_dir.mkdir() + script_file = test_dir / "script.py" + script_file.write_text("print('hello')") + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint=f"python {test_dir}/script.py", # Redundant reference + namespace="test-namespace", + runtime_env=RuntimeEnv(working_dir=str(test_dir)), + ) + + # Should raise ValueError during submit() before API call + with pytest.raises(ValueError, match="Working directory conflict detected"): + rayjob.submit() + + # Verify submit_job was never called (validation blocked it) + mock_api_instance.submit_job.assert_not_called() + + +def test_validate_working_dir_entrypoint_error_message_format( + auto_mock_setup, tmp_path +): + """ + Test that error message contains helpful information. + """ + # Create test directory and file + test_dir = tmp_path / "testdir" + test_dir.mkdir() + script_file = test_dir / "script.py" + script_file.write_text("print('hello')") + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint=f"python {test_dir}/script.py", + namespace="test-namespace", + runtime_env=RuntimeEnv(working_dir=str(test_dir)), + ) + + try: + rayjob._validate_working_dir_entrypoint() + assert False, "Should have raised ValueError" + except ValueError as e: + error_msg = str(e) + # Verify error message contains key information + assert "Working directory conflict detected" in error_msg + assert "working_dir:" in error_msg + assert "entrypoint references:" in error_msg + assert "Fix: Remove the directory prefix" in error_msg + assert "python script.py" in error_msg # Suggested fix + + +def test_validate_working_dir_entrypoint_subdirectory_valid(auto_mock_setup, tmp_path): + """ + Test validation passes when entrypoint references subdirectory within working_dir. + """ + # Create test directory structure: testdir/subdir/script.py + test_dir = tmp_path / "testdir" + test_dir.mkdir() + sub_dir = test_dir / "subdir" + sub_dir.mkdir() + script_file = sub_dir / "script.py" + script_file.write_text("print('hello')") + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint="python subdir/script.py", # Correct: relative to working_dir + namespace="test-namespace", + runtime_env=RuntimeEnv(working_dir=str(test_dir)), + ) + + # Should not raise any exception + rayjob._validate_working_dir_entrypoint() + + +def test_validate_working_dir_entrypoint_runtime_env_as_dict(auto_mock_setup, tmp_path): + """ + Test validation works when runtime_env is passed as dict (not RuntimeEnv object). + """ + # Create test directory and file + test_dir = tmp_path / "testdir" + test_dir.mkdir() + script_file = test_dir / "script.py" + script_file.write_text("print('hello')") + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint=f"python {test_dir}/script.py", # Redundant reference + namespace="test-namespace", + runtime_env={"working_dir": str(test_dir)}, # Dict instead of RuntimeEnv + ) + + # Should raise ValueError even with dict runtime_env + with pytest.raises(ValueError, match="Working directory conflict detected"): + rayjob._validate_working_dir_entrypoint() + + +def test_validate_file_exists_with_working_dir(auto_mock_setup, tmp_path): + """ + Test validation checks that entrypoint file exists within working_dir. + """ + # Create working directory but NOT the script file + test_dir = tmp_path / "testdir" + test_dir.mkdir() + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint="python script.py", # File doesn't exist + namespace="test-namespace", + runtime_env=RuntimeEnv(working_dir=str(test_dir)), + ) + + # Should raise ValueError about missing file + with pytest.raises(ValueError, match="Entrypoint file not found"): + rayjob._validate_working_dir_entrypoint() + + +def test_validate_file_exists_without_working_dir(auto_mock_setup, tmp_path): + """ + Test validation checks that entrypoint file exists when no working_dir and using ./ prefix. + """ + # Don't create the script file + script_path = "./missing_script.py" + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint=f"python {script_path}", # File doesn't exist (local path with ./) + namespace="test-namespace", + ) + + # Should raise ValueError about missing file + with pytest.raises(ValueError, match="Entrypoint file not found"): + rayjob._validate_working_dir_entrypoint() + + +def test_validate_existing_file_with_working_dir_passes(auto_mock_setup, tmp_path): + """ + Test validation passes when file exists in working_dir. + """ + # Create working directory AND the script file + test_dir = tmp_path / "testdir" + test_dir.mkdir() + script_file = test_dir / "script.py" + script_file.write_text("print('hello')") + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint="python script.py", # File exists + namespace="test-namespace", + runtime_env=RuntimeEnv(working_dir=str(test_dir)), + ) + + # Should not raise any exception + rayjob._validate_working_dir_entrypoint() + + +def test_validate_inline_python_command_skipped(auto_mock_setup): + """ + Test validation skips inline Python commands (no file reference). + """ + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint="python -c 'print(\"hello world\")'", # No file reference + namespace="test-namespace", + ) + + # Should not raise any exception (no file to validate) + rayjob._validate_working_dir_entrypoint() + + +def test_validate_simple_filename_without_working_dir_missing(auto_mock_setup): + """ + Test validation checks simple filenames without working_dir. + """ + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint="python script.py", # File doesn't exist locally + namespace="test-namespace", + ) + + # Should raise ValueError (file will be extracted from local, so must exist) + with pytest.raises(ValueError, match="Entrypoint file not found"): + rayjob._validate_working_dir_entrypoint() + + +def test_validate_simple_filename_without_working_dir_exists(auto_mock_setup, tmp_path): + """ + Test validation passes when simple filename exists locally without working_dir. + """ + import os + + original_cwd = os.getcwd() + try: + os.chdir(tmp_path) + # Create the script file in current directory + script_file = tmp_path / "script.py" + script_file.write_text("print('hello')") + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + entrypoint="python script.py", # Simple filename exists locally + namespace="test-namespace", + ) + + # Should not raise exception (file exists) + rayjob._validate_working_dir_entrypoint() + finally: + os.chdir(original_cwd) diff --git a/src/codeflare_sdk/ray/rayjobs/test/test_runtime_env.py b/src/codeflare_sdk/ray/rayjobs/test/test_runtime_env.py new file mode 100644 index 00000000..e059a8d3 --- /dev/null +++ b/src/codeflare_sdk/ray/rayjobs/test/test_runtime_env.py @@ -0,0 +1,1061 @@ +# Copyright 2025 IBM, Red Hat +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + + +import pytest +import os +import io +from unittest.mock import MagicMock, patch +from codeflare_sdk.common.utils.constants import MOUNT_PATH, RAY_VERSION +from ray.runtime_env import RuntimeEnv + +from codeflare_sdk.ray.rayjobs.rayjob import RayJob +from codeflare_sdk.ray.cluster.config import ClusterConfiguration +from codeflare_sdk.ray.rayjobs.config import ManagedClusterConfig +from kubernetes.client import ( + V1Volume, + V1VolumeMount, + ApiException, +) + +from codeflare_sdk.ray.rayjobs.runtime_env import ( + create_secret_from_spec, + extract_all_local_files, +) + + +def test_rayjob_with_remote_working_dir(auto_mock_setup): + """ + Test RayJob with remote working directory in runtime_env. + Should not extract local files and should pass through remote URL. + """ + runtime_env = RuntimeEnv( + working_dir="https://github.com/org/repo/archive/refs/heads/main.zip", + pip=["numpy", "pandas"], + env_vars={"TEST_VAR": "test_value"}, + ) + + rayjob = RayJob( + job_name="test-job", + entrypoint="python test.py", + cluster_name="test-cluster", + runtime_env=runtime_env, + namespace="test-namespace", + ) + + assert rayjob.runtime_env == runtime_env + + # Should not extract any local files due to remote working_dir + files = extract_all_local_files(rayjob) + assert files is None + + rayjob_cr = rayjob._build_rayjob_cr() + + # Should have runtimeEnvYAML with all fields + expected_runtime_env = ( + "env_vars:\n" + " TEST_VAR: test_value\n" + "pip:\n" + "- numpy\n" + "- pandas\n" + "working_dir: https://github.com/org/repo/archive/refs/heads/main.zip\n" + ) + assert rayjob_cr["spec"]["runtimeEnvYAML"] == expected_runtime_env + + # Should not have submitterPodTemplate since no local files + assert "submitterPodTemplate" not in rayjob_cr["spec"] + + # Entrypoint should be unchanged + assert rayjob_cr["spec"]["entrypoint"] == "python test.py" + + +def test_build_file_secret_spec(): + """ + Test building Secret specification for files. + """ + config = ManagedClusterConfig() + files = {"main.py": "print('main')", "helper.py": "print('helper')"} + + spec = config.build_file_secret_spec( + job_name="test-job", namespace="test-namespace", files=files + ) + + assert spec["apiVersion"] == "v1" + assert spec["kind"] == "Secret" + assert spec["type"] == "Opaque" + assert spec["metadata"]["name"] == "test-job-files" + assert spec["metadata"]["namespace"] == "test-namespace" + assert spec["data"] == files + + +def test_build_file_volume_specs(): + """ + Test building volume and mount specifications for files. + """ + config = ManagedClusterConfig() + + volume_spec, mount_spec = config.build_file_volume_specs( + secret_name="test-files", mount_path="/custom/path" + ) + + assert volume_spec["name"] == "ray-job-files" + assert volume_spec["secret"]["secretName"] == "test-files" + + assert mount_spec["name"] == "ray-job-files" + assert mount_spec["mountPath"] == "/custom/path" + + +def test_add_file_volumes(): + """ + Test adding file volumes to cluster configuration. + """ + config = ManagedClusterConfig() + + # Initially no volumes + assert len(config.volumes) == 0 + assert len(config.volume_mounts) == 0 + + config.add_file_volumes(secret_name="test-files") + + assert len(config.volumes) == 1 + assert len(config.volume_mounts) == 1 + + volume = config.volumes[0] + mount = config.volume_mounts[0] + + assert volume.name == "ray-job-files" + assert volume.secret.secret_name == "test-files" + + assert mount.name == "ray-job-files" + assert mount.mount_path == MOUNT_PATH + + +def test_add_file_volumes_duplicate_prevention(): + """ + Test that adding file volumes twice doesn't create duplicates. + """ + config = ManagedClusterConfig() + + # Add volumes twice + config.add_file_volumes(secret_name="test-files") + config.add_file_volumes(secret_name="test-files") + + assert len(config.volumes) == 1 + assert len(config.volume_mounts) == 1 + + +def test_create_secret_from_spec(auto_mock_setup): + """ + Test creating Secret via Kubernetes API. + """ + mock_api_instance = auto_mock_setup["k8s_api"] + + rayjob = RayJob( + job_name="test-job", + cluster_name="existing-cluster", + entrypoint="python test.py", + namespace="test-namespace", + ) + + secret_spec = { + "apiVersion": "v1", + "kind": "Secret", + "type": "Opaque", + "metadata": {"name": "test-files", "namespace": "test-namespace"}, + "data": {"test.py": "print('test')"}, + } + + # Provide valid RayJob result with UID as KubeRay client would + rayjob_result = { + "metadata": { + "name": "test-job", + "namespace": "test-namespace", + "uid": "test-uid-12345", + } + } + + result = create_secret_from_spec(rayjob, secret_spec, rayjob_result) + + assert result == "test-files" + mock_api_instance.create_namespaced_secret.assert_called_once() + + +def test_create_secret_already_exists(auto_mock_setup): + """ + Test creating Secret when it already exists (409 conflict). + """ + mock_api_instance = auto_mock_setup["k8s_api"] + + mock_api_instance.create_namespaced_secret.side_effect = ApiException(status=409) + + rayjob = RayJob( + job_name="test-job", + cluster_name="existing-cluster", + entrypoint="python test.py", + namespace="test-namespace", + ) + + secret_spec = { + "apiVersion": "v1", + "kind": "Secret", + "type": "Opaque", + "metadata": {"name": "test-files", "namespace": "test-namespace"}, + "data": {"test.py": "print('test')"}, + } + + # Provide valid RayJob result with UID as KubeRay client would + rayjob_result = { + "metadata": { + "name": "test-job", + "namespace": "test-namespace", + "uid": "test-uid-67890", + } + } + + result = create_secret_from_spec(rayjob, secret_spec, rayjob_result) + + assert result == "test-files" + mock_api_instance.create_namespaced_secret.assert_called_once() + mock_api_instance.replace_namespaced_secret.assert_called_once() + + +def test_create_secret_with_owner_reference_basic(mocker, auto_mock_setup, caplog): + """ + Test creating Secret with owner reference from valid RayJob result. + """ + mock_api_instance = auto_mock_setup["k8s_api"] + + # Mock client.V1ObjectMeta and V1Secret + mock_v1_metadata = mocker.patch("kubernetes.client.V1ObjectMeta") + mock_metadata_instance = MagicMock() + mock_v1_metadata.return_value = mock_metadata_instance + + rayjob = RayJob( + job_name="test-job", + cluster_name="existing-cluster", + entrypoint="python test.py", + namespace="test-namespace", + ) + + secret_spec = { + "apiVersion": "v1", + "kind": "Secret", + "type": "Opaque", + "metadata": { + "name": "test-files", + "namespace": "test-namespace", + "labels": { + "ray.io/job-name": "test-job", + "app.kubernetes.io/managed-by": "codeflare-sdk", + "app.kubernetes.io/component": "rayjob-files", + }, + }, + "data": {"test.py": "print('test')"}, + } + + # Valid RayJob result with UID + rayjob_result = { + "metadata": { + "name": "test-job", + "namespace": "test-namespace", + "uid": "a4dd4c5a-ab61-411d-b4d1-4abb5177422a", + } + } + + with caplog.at_level("INFO"): + result = create_secret_from_spec(rayjob, secret_spec, rayjob_result) + + assert result == "test-files" + + # Verify owner reference was set + expected_owner_ref = mocker.ANY # We'll check via the logs + assert ( + "Adding owner reference to Secret 'test-files' with RayJob UID: a4dd4c5a-ab61-411d-b4d1-4abb5177422a" + in caplog.text + ) + + assert mock_metadata_instance.owner_references is not None + mock_api_instance.create_namespaced_secret.assert_called_once() + + +def test_file_handling_kubernetes_best_practice_flow(mocker, tmp_path): + """ + Test the Kubernetes best practice flow: pre-declare volume, submit, create Secret. + """ + mocker.patch("kubernetes.config.load_kube_config") + + mock_api_class = mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayjobApi") + mock_api_instance = MagicMock() + mock_api_class.return_value = mock_api_instance + + submit_result = { + "metadata": { + "name": "test-job", + "namespace": "test-namespace", + "uid": "test-uid-12345", + } + } + mock_api_instance.submit_job.return_value = submit_result + + # Mock create_file_secret where it's used (imported into rayjob module) + mock_create_secret = mocker.patch( + "codeflare_sdk.ray.rayjobs.rayjob.create_file_secret" + ) + mock_add_volumes = mocker.patch.object(ManagedClusterConfig, "add_file_volumes") + + # RayClusterApi is already mocked by auto_mock_setup + + test_file = tmp_path / "test.py" + test_file.write_text("print('test')") + + call_order = [] + + def track_add_volumes(*args, **kwargs): + call_order.append("add_volumes") + # Should be called with Secret name + assert args[0] == "test-job-files" + + def track_submit(*args, **kwargs): + call_order.append("submit_job") + return submit_result + + def track_create_secret(*args, **kwargs): + call_order.append("create_secret") + # Args should be: (job, files, rayjob_result) + assert len(args) >= 3, f"Expected 3 args, got {len(args)}: {args}" + assert args[2] == submit_result # rayjob_result should be third arg + + mock_add_volumes.side_effect = track_add_volumes + mock_api_instance.submit_job.side_effect = track_submit + mock_create_secret.side_effect = track_create_secret + + original_cwd = os.getcwd() + try: + os.chdir(tmp_path) + + cluster_config = ManagedClusterConfig() + + rayjob = RayJob( + job_name="test-job", + cluster_config=cluster_config, + entrypoint="python test.py", + namespace="test-namespace", + ) + + rayjob.submit() + finally: + os.chdir(original_cwd) + + # Verify the order: submit → create Secret + assert call_order == ["submit_job", "create_secret"] + + mock_api_instance.submit_job.assert_called_once() + mock_create_secret.assert_called_once() + + # Verify create_file_secret was called with: (job, files, rayjob_result) + # Files dict includes metadata key __entrypoint_path__ for single file case + call_args = mock_create_secret.call_args[0] + assert call_args[0] == rayjob + assert call_args[2] == submit_result + # Check that the actual file content is present + assert "test.py" in call_args[1] + assert call_args[1]["test.py"] == "print('test')" + + +def test_rayjob_submit_with_files_new_cluster(auto_mock_setup, tmp_path): + """ + Test RayJob submission with file detection for new cluster. + """ + mock_api_instance = auto_mock_setup["rayjob_api"] + mock_api_instance.submit_job.return_value = { + "metadata": { + "name": "test-job", + "namespace": "test-namespace", + "uid": "test-uid-files-12345", + } + } + + mock_k8s_instance = auto_mock_setup["k8s_api"] + + # Create test file + test_file = tmp_path / "test.py" + test_file.write_text("print('Hello from the test file!')") + + cluster_config = ManagedClusterConfig() + + original_cwd = os.getcwd() + os.chdir(tmp_path) + + try: + rayjob = RayJob( + job_name="test-job", + cluster_config=cluster_config, + entrypoint="python test.py", + namespace="test-namespace", + ) + + # Submit should detect files and handle them + result = rayjob.submit() + + assert result == "test-job" + + mock_k8s_instance.create_namespaced_secret.assert_called_once() + + assert len(cluster_config.volumes) == 0 + assert len(cluster_config.volume_mounts) == 0 + # Entrypoint should be adjusted to use just the filename + assert rayjob.entrypoint == "python test.py" + + finally: + os.chdir(original_cwd) + + +def test_create_secret_api_error_non_409(auto_mock_setup): + """ + Test create_secret_from_spec handles non-409 API errors. + """ + mock_api_instance = auto_mock_setup["k8s_api"] + + # Configure to raise 500 error + mock_api_instance.create_namespaced_secret.side_effect = ApiException(status=500) + + rayjob = RayJob( + job_name="test-job", + cluster_name="existing-cluster", + entrypoint="python test.py", + namespace="test-namespace", + ) + + secret_spec = { + "apiVersion": "v1", + "kind": "Secret", + "type": "Opaque", + "metadata": {"name": "test-files", "namespace": "test-namespace"}, + "data": {"test.py": "print('test')"}, + } + + # Provide valid RayJob result with UID as KubeRay client would + rayjob_result = { + "metadata": { + "name": "test-job", + "namespace": "test-namespace", + "uid": "test-uid-api-error", + } + } + + with pytest.raises(RuntimeError, match="Failed to create Secret"): + create_secret_from_spec(rayjob, secret_spec, rayjob_result) + + +def test_add_file_volumes_existing_volume_skip(): + """ + Test add_file_volumes skips when volume already exists (missing coverage). + """ + from kubernetes.client import V1SecretVolumeSource + + config = ManagedClusterConfig() + + # Pre-add a volume with same name + existing_volume = V1Volume( + name="ray-job-files", + secret=V1SecretVolumeSource(secret_name="existing-files"), + ) + config.volumes.append(existing_volume) + + config.add_file_volumes(secret_name="new-files") + assert len(config.volumes) == 1 + assert len(config.volume_mounts) == 0 # Mount not added due to volume skip + + +def test_add_file_volumes_existing_mount_skip(): + """ + Test add_file_volumes skips when mount already exists (missing coverage). + """ + config = ManagedClusterConfig() + + # Pre-add a mount with same name + existing_mount = V1VolumeMount(name="ray-job-files", mount_path="/existing/path") + config.volume_mounts.append(existing_mount) + + config.add_file_volumes(secret_name="new-files") + assert len(config.volumes) == 0 # Volume not added due to mount skip + assert len(config.volume_mounts) == 1 + + +def test_zip_directory_functionality(tmp_path): + """ + Test _zip_directory with real directories and files. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import _zip_directory + + # Create test directory structure + test_dir = tmp_path / "working_dir" + test_dir.mkdir() + + # Create some test files + (test_dir / "main.py").write_text("print('main script')") + (test_dir / "utils.py").write_text("def helper(): pass") + + # Create subdirectory with file + sub_dir = test_dir / "subdir" + sub_dir.mkdir() + (sub_dir / "nested.py").write_text("print('nested file')") + + # Test zipping + zip_data = _zip_directory(str(test_dir)) + + assert zip_data is not None + assert len(zip_data) > 0 + assert isinstance(zip_data, bytes) + + +def test_zip_directory_excludes_jupyter_notebooks(tmp_path, caplog): + """ + Test that Jupyter notebook files (.ipynb) and markdown files (.md) are excluded from zip. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import _zip_directory + import zipfile + + # Create test directory with mixed file types + test_dir = tmp_path / "working_dir" + test_dir.mkdir() + + # Create Python files (should be included) + (test_dir / "main.py").write_text("print('main script')") + (test_dir / "utils.py").write_text("def helper(): pass") + + # Create Jupyter notebook files (should be excluded) + (test_dir / "analysis.ipynb").write_text('{"cells": [], "metadata": {}}') + (test_dir / "experiment.IPYNB").write_text( + '{"cells": [], "metadata": {}}' + ) # Test case insensitive + + # Create markdown files (should be excluded) + (test_dir / "README.md").write_text("# Project Documentation\n") + (test_dir / "CHANGELOG.MD").write_text("# Changes\n") # Test case insensitive + + # Create subdirectory with mixed files + sub_dir = test_dir / "notebooks" + sub_dir.mkdir() + (sub_dir / "data_exploration.ipynb").write_text('{"cells": [], "metadata": {}}') + (sub_dir / "helper.py").write_text("print('nested file')") + (sub_dir / "guide.md").write_text("# Guide\n") + + # Test zipping + with caplog.at_level("INFO"): + zip_data = _zip_directory(str(test_dir)) + + assert zip_data is not None + assert len(zip_data) > 0 + + # Verify log message includes exclusion count (3 ipynb + 3 md = 6 total) + assert "Excluded 6 file(s) (.ipynb, .md)" in caplog.text + + # Verify excluded files are not in the zip + zip_buffer = io.BytesIO(zip_data) + with zipfile.ZipFile(zip_buffer, "r") as zipf: + zip_contents = zipf.namelist() + + # Python files should be present + assert "main.py" in zip_contents + assert "utils.py" in zip_contents + assert "notebooks/helper.py" in zip_contents + + # Jupyter notebooks should be excluded + assert "analysis.ipynb" not in zip_contents + assert "experiment.IPYNB" not in zip_contents + assert "notebooks/data_exploration.ipynb" not in zip_contents + + # Markdown files should be excluded + assert "README.md" not in zip_contents + assert "CHANGELOG.MD" not in zip_contents + assert "notebooks/guide.md" not in zip_contents + + +def test_zip_directory_no_exclusions_when_no_notebooks(tmp_path, caplog): + """ + Test that no exclusion message is logged when no notebook or markdown files exist. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import _zip_directory + + # Create test directory with only Python files + test_dir = tmp_path / "working_dir" + test_dir.mkdir() + (test_dir / "main.py").write_text("print('main script')") + (test_dir / "utils.py").write_text("def helper(): pass") + + # Test zipping + with caplog.at_level("INFO"): + zip_data = _zip_directory(str(test_dir)) + + assert zip_data is not None + + # Verify log message does NOT mention exclusions + assert "Excluded" not in caplog.text + + +def test_should_exclude_file_function(): + """ + Test the _should_exclude_file helper function directly. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import _should_exclude_file + + # Should exclude .ipynb files (case insensitive) + assert _should_exclude_file("notebook.ipynb") is True + assert _should_exclude_file("analysis.IPYNB") is True + assert _should_exclude_file("data/exploration.ipynb") is True + assert _should_exclude_file("subdir/nested.Ipynb") is True + + # Should exclude .md files (case insensitive) + assert _should_exclude_file("README.md") is True + assert _should_exclude_file("CHANGELOG.MD") is True + assert _should_exclude_file("docs/guide.md") is True + assert _should_exclude_file("subdir/notes.Md") is True + + # Should NOT exclude other files + assert _should_exclude_file("script.py") is False + assert _should_exclude_file("data.json") is False + assert _should_exclude_file("requirements.txt") is False + assert _should_exclude_file("model.pkl") is False + assert _should_exclude_file("markdown_parser.py") is False # Not .md + assert _should_exclude_file("test.html") is False + + +def test_zip_directory_error_handling(): + """ + Test _zip_directory error handling for IO errors during zipping. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import _zip_directory + + # Mock os.walk to raise an OSError + with patch("os.walk", side_effect=OSError("Permission denied")): + zip_data = _zip_directory("/some/path") + assert zip_data is None + + +def test_extract_all_local_files_with_working_dir(tmp_path): + """ + Test extract_all_local_files with local working directory. + """ + # Create test working directory + working_dir = tmp_path / "working_dir" + working_dir.mkdir() + (working_dir / "script.py").write_text("print('working dir script')") + + runtime_env = RuntimeEnv(working_dir=str(working_dir)) + + rayjob = RayJob( + job_name="test-job", + entrypoint="python script.py", + runtime_env=runtime_env, + namespace="test-namespace", + cluster_name="test-cluster", + ) + + files = extract_all_local_files(rayjob) + + assert files is not None + assert "working_dir.zip" in files + assert isinstance(files["working_dir.zip"], str) # base64 encoded + + # Verify it's valid base64 + import base64 + + try: + decoded = base64.b64decode(files["working_dir.zip"]) + assert len(decoded) > 0 + except Exception: + pytest.fail("Invalid base64 encoding") + + +def test_extract_all_local_files_excludes_notebooks(tmp_path, caplog): + """ + Test that extract_all_local_files excludes Jupyter notebooks and markdown files when zipping working directory. + """ + import zipfile + import base64 + + # Create test working directory with mixed files + working_dir = tmp_path / "working_dir" + working_dir.mkdir() + + # Python files that should be included + (working_dir / "main.py").write_text("print('main script')") + (working_dir / "helper.py").write_text("def helper_function(): pass") + + # Jupyter notebooks that should be excluded + (working_dir / "analysis.ipynb").write_text( + '{"cells": [{"cell_type": "code", "source": ["print(\'hello\')"]}]}' + ) + (working_dir / "data.ipynb").write_text('{"cells": [], "metadata": {}}') + + # Markdown files that should be excluded + (working_dir / "README.md").write_text("# Project Documentation\n") + (working_dir / "CHANGELOG.md").write_text("# Changes\n") + + runtime_env = RuntimeEnv(working_dir=str(working_dir)) + + rayjob = RayJob( + job_name="test-job", + entrypoint="python main.py", + runtime_env=runtime_env, + namespace="test-namespace", + cluster_name="test-cluster", + ) + + # This should zip the directory and exclude notebooks and markdown files + with caplog.at_level("INFO"): + files = extract_all_local_files(rayjob) + + assert files is not None + assert "working_dir.zip" in files + + # Verify exclusion was logged (2 ipynb + 2 md = 4 total) + assert "Excluded 4 file(s) (.ipynb, .md)" in caplog.text + + # Decode and verify zip contents + zip_data = base64.b64decode(files["working_dir.zip"]) + zip_buffer = io.BytesIO(zip_data) + + with zipfile.ZipFile(zip_buffer, "r") as zipf: + zip_contents = zipf.namelist() + + # Python files should be present + assert "main.py" in zip_contents + assert "helper.py" in zip_contents + + # Jupyter notebooks should be excluded + assert "analysis.ipynb" not in zip_contents + assert "data.ipynb" not in zip_contents + + # Markdown files should be excluded + assert "README.md" not in zip_contents + assert "CHANGELOG.md" not in zip_contents + + +def test_extract_single_entrypoint_file_error_handling(tmp_path): + """ + Test _extract_single_entrypoint_file with file read errors. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import _extract_single_entrypoint_file + + # Create a file that exists but make it unreadable + test_file = tmp_path / "unreadable.py" + test_file.write_text("print('test')") + + rayjob = RayJob( + job_name="test-job", + entrypoint=f"python {test_file}", + namespace="test-namespace", + cluster_name="test-cluster", + ) + + # Mock open to raise IOError + with patch("builtins.open", side_effect=IOError("Permission denied")): + result = _extract_single_entrypoint_file(rayjob) + assert result is None + + +def test_extract_single_entrypoint_file_no_match(): + """ + Test _extract_single_entrypoint_file with no Python file matches. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import _extract_single_entrypoint_file + + rayjob = RayJob( + job_name="test-job", + entrypoint="echo 'no python files here'", + namespace="test-namespace", + cluster_name="test-cluster", + ) + + result = _extract_single_entrypoint_file(rayjob) + assert result is None + + +def test_parse_requirements_file_valid(tmp_path): + """ + Test parse_requirements_file with valid requirements.txt. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import parse_requirements_file + + # Create test requirements file + req_file = tmp_path / "requirements.txt" + req_file.write_text( + """# This is a comment +numpy==1.21.0 +pandas>=1.3.0 + +# Another comment +scikit-learn +""" + ) + + result = parse_requirements_file(str(req_file)) + + assert result is not None + assert len(result) == 3 + assert "numpy==1.21.0" in result + assert "pandas>=1.3.0" in result + assert "scikit-learn" in result + # Comments and empty lines should be filtered out + assert "# This is a comment" not in result + + +def test_parse_requirements_file_missing(): + """ + Test parse_requirements_file with non-existent file. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import parse_requirements_file + + result = parse_requirements_file("/non/existent/requirements.txt") + assert result is None + + +def test_parse_requirements_file_read_error(tmp_path): + """ + Test parse_requirements_file with file read error. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import parse_requirements_file + + # Create a file + req_file = tmp_path / "requirements.txt" + req_file.write_text("numpy==1.21.0") + + # Mock open to raise IOError + with patch("builtins.open", side_effect=IOError("Permission denied")): + result = parse_requirements_file(str(req_file)) + assert result is None + + +def test_process_pip_dependencies_list(): + """ + Test process_pip_dependencies with list input. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import process_pip_dependencies + + rayjob = RayJob( + job_name="test-job", + entrypoint="python test.py", + namespace="test-namespace", + cluster_name="test-cluster", + ) + + pip_list = ["numpy", "pandas", "scikit-learn"] + result = process_pip_dependencies(rayjob, pip_list) + + assert result == pip_list + + +def test_process_pip_dependencies_requirements_file(tmp_path): + """ + Test process_pip_dependencies with requirements.txt path. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import process_pip_dependencies + + # Create test requirements file + req_file = tmp_path / "requirements.txt" + req_file.write_text("numpy==1.21.0\npandas>=1.3.0") + + rayjob = RayJob( + job_name="test-job", + entrypoint="python test.py", + namespace="test-namespace", + cluster_name="test-cluster", + ) + + result = process_pip_dependencies(rayjob, str(req_file)) + + assert result is not None + assert len(result) == 2 + assert "numpy==1.21.0" in result + assert "pandas>=1.3.0" in result + + +def test_process_pip_dependencies_dict_format(): + """ + Test process_pip_dependencies with dict format containing packages. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import process_pip_dependencies + + rayjob = RayJob( + job_name="test-job", + entrypoint="python test.py", + namespace="test-namespace", + cluster_name="test-cluster", + ) + + pip_dict = {"packages": ["numpy", "pandas"], "pip_check": False} + result = process_pip_dependencies(rayjob, pip_dict) + + assert result == ["numpy", "pandas"] + + +def test_process_pip_dependencies_unsupported_format(): + """ + Test process_pip_dependencies with unsupported format. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import process_pip_dependencies + + rayjob = RayJob( + job_name="test-job", + entrypoint="python test.py", + namespace="test-namespace", + cluster_name="test-cluster", + ) + + # Test with unsupported format (int) + result = process_pip_dependencies(rayjob, 12345) + assert result is None + + +def test_process_runtime_env_local_working_dir(tmp_path): + """ + Test process_runtime_env with local working directory. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import process_runtime_env, UNZIP_PATH + + # Create test working directory + working_dir = tmp_path / "working_dir" + working_dir.mkdir() + (working_dir / "script.py").write_text("print('test')") + + runtime_env = RuntimeEnv( + working_dir=str(working_dir), + env_vars={"TEST_VAR": "test_value"}, + ) + + rayjob = RayJob( + job_name="test-job", + entrypoint="python script.py", + runtime_env=runtime_env, + namespace="test-namespace", + cluster_name="test-cluster", + ) + + result = process_runtime_env(rayjob) + + assert result is not None + assert f"working_dir: {UNZIP_PATH}" in result + assert "env_vars:" in result + assert "TEST_VAR: test_value" in result + + +def test_process_runtime_env_single_file_case(tmp_path): + """ + Test process_runtime_env with single file case (no working_dir in runtime_env). + """ + from codeflare_sdk.ray.rayjobs.runtime_env import process_runtime_env + from codeflare_sdk.common.utils.constants import MOUNT_PATH + + rayjob = RayJob( + job_name="test-job", + entrypoint="python test.py", + namespace="test-namespace", + cluster_name="test-cluster", + ) + + # Files dict without working_dir.zip (single file case) + files = {"test.py": "print('test')"} + + result = process_runtime_env(rayjob, files) + + assert result is not None + assert f"working_dir: {MOUNT_PATH}" in result + + +def test_process_runtime_env_no_processing_needed(): + """ + Test process_runtime_env returns None when no processing needed. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import process_runtime_env + + rayjob = RayJob( + job_name="test-job", + entrypoint="python test.py", + namespace="test-namespace", + cluster_name="test-cluster", + ) + + # No runtime_env and no files + result = process_runtime_env(rayjob) + assert result is None + + +def test_normalize_runtime_env_with_none(): + """ + Test _normalize_runtime_env with None input. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import _normalize_runtime_env + + result = _normalize_runtime_env(None) + assert result is None + + +def test_extract_single_entrypoint_file_no_entrypoint(): + """ + Test _extract_single_entrypoint_file with no entrypoint. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import _extract_single_entrypoint_file + + rayjob = RayJob( + job_name="test-job", + entrypoint=None, # No entrypoint + namespace="test-namespace", + cluster_name="test-cluster", + ) + + result = _extract_single_entrypoint_file(rayjob) + assert result is None + + +def test_create_file_secret_filters_metadata_keys(auto_mock_setup, tmp_path): + """ + Test create_file_secret filters out metadata keys from files dict. + """ + from codeflare_sdk.ray.rayjobs.runtime_env import create_file_secret + + rayjob = RayJob( + job_name="test-job", + cluster_name="existing-cluster", + entrypoint="python test.py", + namespace="test-namespace", + ) + + # Files dict with metadata key that should be filtered out + files = { + "__entrypoint_path__": "some/path/test.py", # Should be filtered + "test.py": "print('test')", # Should remain + } + + rayjob_result = { + "metadata": { + "name": "test-job", + "namespace": "test-namespace", + "uid": "test-uid-12345", + } + } + + # This should not raise an error and should filter out metadata keys + create_file_secret(rayjob, files, rayjob_result) + + # Verify the Secret was created (mocked) + mock_api_instance = auto_mock_setup["k8s_api"] + mock_api_instance.create_namespaced_secret.assert_called_once() + + # The call should have filtered data (only test.py, not __entrypoint_path__) + call_args = mock_api_instance.create_namespaced_secret.call_args + secret_data = call_args[1]["body"].string_data # Changed from data to string_data + assert "test.py" in secret_data + assert "__entrypoint_path__" not in secret_data diff --git a/src/codeflare_sdk/ray/rayjobs/test/test_status.py b/src/codeflare_sdk/ray/rayjobs/test/test_status.py new file mode 100644 index 00000000..2f2b9957 --- /dev/null +++ b/src/codeflare_sdk/ray/rayjobs/test/test_status.py @@ -0,0 +1,373 @@ +# Copyright 2025 IBM, Red Hat +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +from codeflare_sdk.ray.rayjobs.rayjob import RayJob +from codeflare_sdk.ray.rayjobs.status import ( + CodeflareRayJobStatus, + RayJobDeploymentStatus, + RayJobInfo, +) + + +def test_rayjob_status(mocker): + """ + Test the RayJob status method with different deployment statuses. + """ + # Mock kubernetes config loading + mocker.patch("kubernetes.config.load_kube_config") + # Mock the RayjobApi to avoid actual Kubernetes calls + mock_api_class = mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayjobApi") + mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayClusterApi") + mock_api_instance = mock_api_class.return_value + + # Create a RayJob instance + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + namespace="test-ns", + entrypoint="python test.py", + ) + + # Test case 1: No job found + mock_api_instance.get_job_status.return_value = None + status, ready = rayjob.status(print_to_console=False) + assert status == CodeflareRayJobStatus.UNKNOWN + assert ready == False + + # Test case 2: Running job + mock_api_instance.get_job_status.return_value = { + "jobId": "test-job-abc123", + "jobDeploymentStatus": "Running", + "startTime": "2025-07-28T11:37:07Z", + "failed": 0, + "succeeded": 0, + "rayClusterName": "test-cluster", + } + status, ready = rayjob.status(print_to_console=False) + assert status == CodeflareRayJobStatus.RUNNING + assert ready == False + + # Test case 3: Complete job + mock_api_instance.get_job_status.return_value = { + "jobId": "test-job-abc123", + "jobDeploymentStatus": "Complete", + "startTime": "2025-07-28T11:37:07Z", + "endTime": "2025-07-28T11:42:30Z", + "failed": 0, + "succeeded": 1, + "rayClusterName": "test-cluster", + } + status, ready = rayjob.status(print_to_console=False) + assert status == CodeflareRayJobStatus.COMPLETE + assert ready == True + + # Test case 4: Failed job + mock_api_instance.get_job_status.return_value = { + "jobId": "test-job-abc123", + "jobDeploymentStatus": "Failed", + "startTime": "2025-07-28T11:37:07Z", + "endTime": "2025-07-28T11:42:30Z", + "failed": 1, + "succeeded": 0, + "rayClusterName": "test-cluster", + } + status, ready = rayjob.status(print_to_console=False) + assert status == CodeflareRayJobStatus.FAILED + assert ready == False + + # Test case 5: Suspended job + mock_api_instance.get_job_status.return_value = { + "jobId": "test-job-abc123", + "jobDeploymentStatus": "Suspended", + "startTime": "2025-07-28T11:37:07Z", + "failed": 0, + "succeeded": 0, + "rayClusterName": "test-cluster", + } + status, ready = rayjob.status(print_to_console=False) + assert status == CodeflareRayJobStatus.SUSPENDED + assert ready == False + + +def test_rayjob_status_unknown_deployment_status(mocker): + """ + Test handling of unknown deployment status from the API. + """ + mocker.patch("kubernetes.config.load_kube_config") + mock_api_class = mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayjobApi") + mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayClusterApi") + mock_api_instance = mock_api_class.return_value + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + namespace="test-ns", + entrypoint="python test.py", + ) + + # Test with unrecognized deployment status + mock_api_instance.get_job_status.return_value = { + "jobId": "test-job-abc123", + "jobDeploymentStatus": "SomeNewStatus", # Unknown status + "startTime": "2025-07-28T11:37:07Z", + "failed": 0, + "succeeded": 0, + } + + status, ready = rayjob.status(print_to_console=False) + assert status == CodeflareRayJobStatus.UNKNOWN + assert ready == False + + +def test_rayjob_status_missing_fields(mocker): + """ + Test handling of API response with missing fields. + """ + mocker.patch("kubernetes.config.load_kube_config") + mock_api_class = mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayjobApi") + mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayClusterApi") + mock_api_instance = mock_api_class.return_value + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + namespace="test-ns", + entrypoint="python test.py", + ) + + # Test with minimal API response (missing some fields) + mock_api_instance.get_job_status.return_value = { + # Missing jobId, failed, succeeded, etc. + "jobDeploymentStatus": "Running", + } + + status, ready = rayjob.status(print_to_console=False) + assert status == CodeflareRayJobStatus.RUNNING + assert ready == False + + +def test_map_to_codeflare_status(mocker): + """ + Test the _map_to_codeflare_status helper method directly. + """ + # Mock kubernetes config loading + mocker.patch("kubernetes.config.load_kube_config") + # Mock the RayjobApi constructor to avoid authentication issues + mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayjobApi") + mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayClusterApi") + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + namespace="test-ns", + entrypoint="python test.py", + ) + + # Test all deployment status mappings + status, ready = rayjob._map_to_codeflare_status(RayJobDeploymentStatus.COMPLETE) + assert status == CodeflareRayJobStatus.COMPLETE + assert ready == True + + status, ready = rayjob._map_to_codeflare_status(RayJobDeploymentStatus.RUNNING) + assert status == CodeflareRayJobStatus.RUNNING + assert ready == False + + status, ready = rayjob._map_to_codeflare_status(RayJobDeploymentStatus.FAILED) + assert status == CodeflareRayJobStatus.FAILED + assert ready == False + + status, ready = rayjob._map_to_codeflare_status(RayJobDeploymentStatus.SUSPENDED) + assert status == CodeflareRayJobStatus.SUSPENDED + assert ready == False + + status, ready = rayjob._map_to_codeflare_status(RayJobDeploymentStatus.UNKNOWN) + assert status == CodeflareRayJobStatus.UNKNOWN + assert ready == False + + +def test_rayjob_info_dataclass(): + """ + Test the RayJobInfo dataclass creation and field access. + """ + job_info = RayJobInfo( + name="test-job", + job_id="test-job-abc123", + status=RayJobDeploymentStatus.RUNNING, + namespace="test-ns", + cluster_name="test-cluster", + start_time="2025-07-28T11:37:07Z", + failed_attempts=0, + succeeded_attempts=0, + ) + + # Test all fields are accessible + assert job_info.name == "test-job" + assert job_info.job_id == "test-job-abc123" + assert job_info.status == RayJobDeploymentStatus.RUNNING + assert job_info.namespace == "test-ns" + assert job_info.cluster_name == "test-cluster" + assert job_info.start_time == "2025-07-28T11:37:07Z" + assert job_info.end_time is None # Default value + assert job_info.failed_attempts == 0 + assert job_info.succeeded_attempts == 0 + + +def test_rayjob_status_print_no_job_found(mocker): + """ + Test that pretty_print.print_no_job_found is called when no job is found and print_to_console=True. + """ + mocker.patch("kubernetes.config.load_kube_config") + # Mock the RayjobApi and pretty_print + mock_api_class = mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayjobApi") + mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayClusterApi") + mock_api_instance = mock_api_class.return_value + mock_print_no_job_found = mocker.patch( + "codeflare_sdk.ray.rayjobs.pretty_print.print_no_job_found" + ) + + # Create a RayJob instance + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + namespace="test-ns", + entrypoint="python test.py", + ) + + # No job found scenario + mock_api_instance.get_job_status.return_value = None + + # Call status with print_to_console=True + status, ready = rayjob.status(print_to_console=True) + + # Verify the pretty print function was called + mock_print_no_job_found.assert_called_once_with("test-job", "test-ns") + assert status == CodeflareRayJobStatus.UNKNOWN + assert ready == False + + +def test_rayjob_status_print_job_found(mocker): + """ + Test that pretty_print.print_job_status is called when job is found and print_to_console=True. + """ + mocker.patch("kubernetes.config.load_kube_config") + # Mock the RayjobApi and pretty_print + mock_api_class = mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayjobApi") + mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayClusterApi") + mock_api_instance = mock_api_class.return_value + mock_print_job_status = mocker.patch( + "codeflare_sdk.ray.rayjobs.pretty_print.print_job_status" + ) + + # Create a RayJob instance + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + namespace="test-ns", + entrypoint="python test.py", + ) + + # Job found scenario + mock_api_instance.get_job_status.return_value = { + "jobId": "test-job-abc123", + "jobDeploymentStatus": "Running", + "startTime": "2025-07-28T11:37:07Z", + "failed": 0, + "succeeded": 0, + "rayClusterName": "test-cluster", + } + + # Call status with print_to_console=True + status, ready = rayjob.status(print_to_console=True) + + # Verify the pretty print function was called + mock_print_job_status.assert_called_once() + # Verify the RayJobInfo object passed to print_job_status + call_args = mock_print_job_status.call_args[0][0] # First positional argument + assert call_args.name == "test-job" + assert call_args.job_id == "test-job-abc123" + assert call_args.status == RayJobDeploymentStatus.RUNNING + assert call_args.namespace == "test-ns" + assert call_args.cluster_name == "test-cluster" + + assert status == CodeflareRayJobStatus.RUNNING + assert ready == False + + +def test_rayjob_status_all_deployment_states(mocker): + """Test RayJob status method with all deployment states.""" + mocker.patch("kubernetes.config.load_kube_config") + mock_api_class = mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayjobApi") + mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayClusterApi") + mock_api_instance = mock_api_class.return_value + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + namespace="test-ns", + entrypoint="python test.py", + ) + + # Test mapping of deployment statuses to CodeflareRayJobStatus + test_cases = [ + # (deployment_status_str, expected CodeflareRayJobStatus, expected ready) + ("Complete", CodeflareRayJobStatus.COMPLETE, True), + ("Running", CodeflareRayJobStatus.RUNNING, False), + ("Failed", CodeflareRayJobStatus.FAILED, False), + ("Suspended", CodeflareRayJobStatus.SUSPENDED, False), + ] + + for deployment_status_str, expected_status, expected_ready in test_cases: + mock_api_instance.get_job_status.return_value = { + "jobId": "test-job-abc123", + "jobDeploymentStatus": deployment_status_str, + "startTime": "2025-07-28T11:37:07Z", + "failed": 0, + "succeeded": 0, + "rayClusterName": "test-cluster", + } + status, ready = rayjob.status(print_to_console=False) + assert status == expected_status, f"Failed for {deployment_status_str}" + assert ( + ready == expected_ready + ), f"Failed ready check for {deployment_status_str}" + + +def test_rayjob_status_with_end_time(mocker): + """Test RayJob status with end time field.""" + mocker.patch("kubernetes.config.load_kube_config") + mock_api_class = mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayjobApi") + mocker.patch("codeflare_sdk.ray.rayjobs.rayjob.RayClusterApi") + mock_api_instance = mock_api_class.return_value + + rayjob = RayJob( + job_name="test-job", + cluster_name="test-cluster", + namespace="test-ns", + entrypoint="python test.py", + ) + + # Test with end time field + mock_api_instance.get_job_status.return_value = { + "jobId": "test-job-abc123", + "jobDeploymentStatus": "Complete", + "startTime": "2025-07-28T11:37:07Z", + "endTime": "2025-07-28T11:47:07Z", + "failed": 0, + "succeeded": 1, + "rayClusterName": "test-cluster", + } + + status, ready = rayjob.status(print_to_console=False) + assert status == CodeflareRayJobStatus.COMPLETE + assert ready == True diff --git a/tests/e2e/cluster_apply_kind_test.py b/tests/e2e/cluster_apply_kind_test.py index 398bf73b..e023e92d 100644 --- a/tests/e2e/cluster_apply_kind_test.py +++ b/tests/e2e/cluster_apply_kind_test.py @@ -1,6 +1,8 @@ from codeflare_sdk import Cluster, ClusterConfiguration import pytest +import time from kubernetes import client +from codeflare_sdk.common.utils import constants from support import ( initialize_kubernetes_client, @@ -38,6 +40,7 @@ def test_cluster_apply(self): worker_cpu_limits="1", worker_memory_requests="1Gi", worker_memory_limits="2Gi", + image=f"rayproject/ray:{constants.RAY_VERSION}", write_to_file=True, verify_tls=False, ) @@ -47,9 +50,9 @@ def test_cluster_apply(self): cluster.apply() # Wait for the cluster to be ready - cluster.wait_ready() - status = cluster.status() - assert status["ready"], f"Cluster {cluster_name} is not ready: {status}" + cluster.wait_ready(dashboard_check=False) + status, ready = cluster.status() + assert ready, f"Cluster {cluster_name} is not ready: {status}" # Verify the cluster is created ray_cluster = get_ray_cluster(cluster_name, namespace) @@ -58,7 +61,7 @@ def test_cluster_apply(self): ray_cluster["spec"]["workerGroupSpecs"][0]["replicas"] == 1 ), "Initial worker count does not match" - # Update configuration with 3 workers + # Update configuration with 2 workers updated_config = ClusterConfiguration( name=cluster_name, namespace=namespace, @@ -71,6 +74,7 @@ def test_cluster_apply(self): worker_cpu_limits="1", worker_memory_requests="1Gi", worker_memory_limits="2Gi", + image=f"rayproject/ray:{constants.RAY_VERSION}", write_to_file=True, verify_tls=False, ) @@ -79,12 +83,15 @@ def test_cluster_apply(self): cluster.config = updated_config cluster.apply() + # Give Kubernetes a moment to process the update + time.sleep(5) + # Wait for the updated cluster to be ready - cluster.wait_ready() - updated_status = cluster.status() - assert updated_status[ - "ready" - ], f"Cluster {cluster_name} is not ready after update: {updated_status}" + cluster.wait_ready(dashboard_check=False) + updated_status, updated_ready = cluster.status() + assert ( + updated_ready + ), f"Cluster {cluster_name} is not ready after update: {updated_status}" # Verify the cluster is updated updated_ray_cluster = get_ray_cluster(cluster_name, namespace) @@ -94,63 +101,19 @@ def test_cluster_apply(self): # Clean up cluster.down() - ray_cluster = get_ray_cluster(cluster_name, namespace) - assert ray_cluster is None, "Cluster was not deleted successfully" - def test_apply_invalid_update(self): - self.setup_method() - create_namespace(self) + # Wait for deletion to complete (finalizers may delay deletion) + max_wait = 30 # seconds + wait_interval = 2 + elapsed = 0 - cluster_name = "test-cluster-apply-invalid" - namespace = self.namespace + while elapsed < max_wait: + ray_cluster = get_ray_cluster(cluster_name, namespace) + if ray_cluster is None: + break + time.sleep(wait_interval) + elapsed += wait_interval - # Initial configuration - initial_config = ClusterConfiguration( - name=cluster_name, - namespace=namespace, - num_workers=1, - head_cpu_requests="500m", - head_cpu_limits="1", - head_memory_requests="1Gi", - head_memory_limits="2Gi", - worker_cpu_requests="500m", - worker_cpu_limits="1", - worker_memory_requests="1Gi", - worker_memory_limits="2Gi", - write_to_file=True, - verify_tls=False, - ) - - # Create the cluster - cluster = Cluster(initial_config) - cluster.apply() - - # Wait for the cluster to be ready - cluster.wait_ready() - status = cluster.status() - assert status["ready"], f"Cluster {cluster_name} is not ready: {status}" - - # Update with an invalid configuration (e.g., immutable field change) - invalid_config = ClusterConfiguration( - name=cluster_name, - namespace=namespace, - num_workers=2, - head_cpu_requests="1", - head_cpu_limits="2", # Changing CPU limits (immutable) - head_memory_requests="1Gi", - head_memory_limits="2Gi", - worker_cpu_requests="500m", - worker_cpu_limits="1", - worker_memory_requests="1Gi", - worker_memory_limits="2Gi", - write_to_file=True, - verify_tls=False, - ) - - # Try to apply the invalid configuration and expect failure - cluster.config = invalid_config - with pytest.raises(RuntimeError, match="Immutable fields detected"): - cluster.apply() - - # Clean up - cluster.down() + assert ( + ray_cluster is None + ), f"Cluster was not deleted successfully after {max_wait}s" diff --git a/tests/e2e/rayjob/__init__.py b/tests/e2e/rayjob/__init__.py new file mode 100644 index 00000000..e69de29b diff --git a/tests/e2e/rayjob/ray_version_validation_oauth_test.py b/tests/e2e/rayjob/ray_version_validation_oauth_test.py new file mode 100644 index 00000000..794d739a --- /dev/null +++ b/tests/e2e/rayjob/ray_version_validation_oauth_test.py @@ -0,0 +1,127 @@ +import pytest +import sys +import os + +# Add the parent directory to the path to import support +sys.path.insert(0, os.path.join(os.path.dirname(__file__), "..")) +from support import * + +from codeflare_sdk import ( + RayJob, + ManagedClusterConfig, +) + + +class TestRayJobRayVersionValidationOauth: + def setup_method(self): + initialize_kubernetes_client(self) + + def teardown_method(self): + delete_namespace(self) + delete_kueue_resources(self) + + def _create_basic_managed_cluster_config( + self, ray_image: str + ) -> ManagedClusterConfig: + """Helper method to create basic managed cluster configuration.""" + return ManagedClusterConfig( + head_cpu_requests="500m", + head_cpu_limits="500m", + head_memory_requests=1, + head_memory_limits=2, + num_workers=1, + worker_cpu_requests="500m", + worker_cpu_limits="500m", + worker_memory_requests=1, + worker_memory_limits=2, + image=ray_image, + ) + + def test_rayjob_lifecycled_cluster_incompatible_ray_version_oauth(self): + """Test that RayJob creation fails when cluster config specifies incompatible Ray version.""" + self.setup_method() + create_namespace(self) + create_kueue_resources(self) + self.run_rayjob_lifecycled_cluster_incompatible_version() + + def run_rayjob_lifecycled_cluster_incompatible_version(self): + """Test Ray version validation with cluster lifecycling using incompatible image.""" + + job_name = "incompatible-lifecycle-rayjob" + + # Create cluster configuration with incompatible Ray version (2.46.1 instead of expected 2.47.1) + incompatible_ray_image = "quay.io/modh/ray:2.46.1-py311-cu121" + + print( + f"Creating RayJob with incompatible Ray image in cluster config: {incompatible_ray_image}" + ) + + cluster_config = self._create_basic_managed_cluster_config( + incompatible_ray_image + ) + + # Create RayJob with incompatible cluster config - this should fail during submission + rayjob = RayJob( + job_name=job_name, + cluster_config=cluster_config, + namespace=self.namespace, + entrypoint="python -c 'print(\"This should not run due to version mismatch\")'", + ttl_seconds_after_finished=30, + ) + + print( + f"Attempting to submit RayJob '{job_name}' with incompatible Ray version..." + ) + + # This should fail during submission due to Ray version validation + with pytest.raises(ValueError, match="Ray version mismatch detected"): + rayjob.submit() + + print( + "✅ Ray version validation correctly prevented RayJob submission with incompatible cluster config!" + ) + + def test_rayjob_lifecycled_cluster_unknown_ray_version_oauth(self): + """Test that RayJob creation succeeds with warning when Ray version cannot be determined.""" + self.setup_method() + create_namespace(self) + create_kueue_resources(self) + self.run_rayjob_lifecycled_cluster_unknown_version() + + def run_rayjob_lifecycled_cluster_unknown_version(self): + """Test Ray version validation with unknown image (should warn but not fail).""" + + job_name = "unknown-version-rayjob" + + # Use an image where Ray version cannot be determined (SHA digest) + unknown_ray_image = "quay.io/modh/ray@sha256:6d076aeb38ab3c34a6a2ef0f58dc667089aa15826fa08a73273c629333e12f1e" + + print( + f"Creating RayJob with image where Ray version cannot be determined: {unknown_ray_image}" + ) + + cluster_config = self._create_basic_managed_cluster_config(unknown_ray_image) + + # Create RayJob with unknown version image - this should succeed with warning + rayjob = RayJob( + job_name=job_name, + cluster_config=cluster_config, + namespace=self.namespace, + entrypoint="python -c 'print(\"Testing unknown Ray version scenario\")'", + ttl_seconds_after_finished=30, + ) + + print(f"Attempting to submit RayJob '{job_name}' with unknown Ray version...") + + # This should succeed but with a warning + with pytest.warns(UserWarning, match="Cannot determine Ray version"): + submission_result = rayjob.submit() + + assert ( + submission_result == job_name + ), f"Job submission failed, expected {job_name}, got {submission_result}" + + print("✅ RayJob submission succeeded with warning for unknown Ray version!") + print( + f"Note: RayJob '{job_name}' was submitted successfully but may need manual cleanup." + ) diff --git a/tests/e2e/rayjob/rayjob_existing_cluster_test.py b/tests/e2e/rayjob/rayjob_existing_cluster_test.py new file mode 100644 index 00000000..82858d28 --- /dev/null +++ b/tests/e2e/rayjob/rayjob_existing_cluster_test.py @@ -0,0 +1,119 @@ +import pytest +import sys +import os +from time import sleep + +sys.path.insert(0, os.path.join(os.path.dirname(__file__), "..")) +from support import * + +from codeflare_sdk import ( + Cluster, + ClusterConfiguration, +) +from codeflare_sdk import RayJob, TokenAuthentication +from codeflare_sdk.ray.rayjobs.status import CodeflareRayJobStatus +from python_client.kuberay_job_api import RayjobApi + + +class TestRayJobExistingCluster: + """Test RayJob against existing Kueue-managed clusters.""" + + def setup_method(self): + initialize_kubernetes_client(self) + + def teardown_method(self): + delete_namespace(self) + delete_kueue_resources(self) + + def test_existing_kueue_cluster(self): + """Test RayJob against Kueue-managed RayCluster.""" + self.setup_method() + create_namespace(self) + create_kueue_resources(self) + + cluster_name = "kueue-cluster" + + if is_openshift(): + auth = TokenAuthentication( + token=run_oc_command(["whoami", "--show-token=true"]), + server=run_oc_command(["whoami", "--show-server=true"]), + skip_tls=True, + ) + auth.login() + + resources = get_platform_appropriate_resources() + + cluster = Cluster( + ClusterConfiguration( + name=cluster_name, + namespace=self.namespace, + num_workers=1, + head_cpu_requests=resources["head_cpu_requests"], + head_cpu_limits=resources["head_cpu_limits"], + head_memory_requests=resources["head_memory_requests"], + head_memory_limits=resources["head_memory_limits"], + worker_cpu_requests=resources["worker_cpu_requests"], + worker_cpu_limits=resources["worker_cpu_limits"], + worker_memory_requests=resources["worker_memory_requests"], + worker_memory_limits=resources["worker_memory_limits"], + image=constants.CUDA_PY312_RUNTIME_IMAGE, + local_queue=self.local_queues[0], + write_to_file=True, + verify_tls=False, + ) + ) + + cluster.apply() + + # Wait for cluster to be ready (with Kueue admission) + print(f"Waiting for cluster '{cluster_name}' to be ready...") + cluster.wait_ready(timeout=600) + print(f"✓ Cluster '{cluster_name}' is ready") + + # RayJob with explicit local_queue + rayjob_explicit = RayJob( + job_name="job-explicit-queue", + cluster_name=cluster_name, + namespace=self.namespace, + entrypoint="python -c \"import ray; ray.init(); print('Job with explicit queue')\"", + runtime_env={"env_vars": get_setup_env_variables(ACCELERATOR="cpu")}, + local_queue=self.local_queues[0], + ) + + # RayJob using default queue + rayjob_default = RayJob( + job_name="job-default-queue", + cluster_name=cluster_name, + namespace=self.namespace, + entrypoint="python -c \"import ray; ray.init(); print('Job with default queue')\"", + runtime_env={"env_vars": get_setup_env_variables(ACCELERATOR="cpu")}, + ) + + try: + # Test RayJob with explicit queue + assert rayjob_explicit.submit() == "job-explicit-queue" + self._wait_completion(rayjob_explicit) + + # Test RayJob with default queue + assert rayjob_default.submit() == "job-default-queue" + self._wait_completion(rayjob_default) + finally: + rayjob_explicit.delete() + rayjob_default.delete() + cluster.down() + + def _wait_completion(self, rayjob: RayJob, timeout: int = 600): + """Wait for RayJob completion.""" + elapsed = 0 + interval = 10 + + while elapsed < timeout: + status, _ = rayjob.status(print_to_console=False) + if status == CodeflareRayJobStatus.COMPLETE: + return + elif status == CodeflareRayJobStatus.FAILED: + raise AssertionError(f"RayJob '{rayjob.name}' failed") + sleep(interval) + elapsed += interval + + raise TimeoutError(f"RayJob '{rayjob.name}' timeout after {timeout}s") diff --git a/tests/e2e/rayjob/rayjob_lifecycled_cluster_test.py b/tests/e2e/rayjob/rayjob_lifecycled_cluster_test.py new file mode 100644 index 00000000..10390011 --- /dev/null +++ b/tests/e2e/rayjob/rayjob_lifecycled_cluster_test.py @@ -0,0 +1,249 @@ +import pytest +import sys +import os +from time import sleep +import tempfile + +sys.path.insert(0, os.path.join(os.path.dirname(__file__), "..")) +from support import * + +from codeflare_sdk import RayJob, ManagedClusterConfig + +from kubernetes import client +from python_client.kuberay_job_api import RayjobApi +from python_client.kuberay_cluster_api import RayClusterApi + + +class TestRayJobLifecycledCluster: + """Test RayJob with auto-created cluster lifecycle management.""" + + def setup_method(self): + initialize_kubernetes_client(self) + + def teardown_method(self): + delete_namespace(self) + delete_kueue_resources(self) + + def test_lifecycled_kueue_managed(self): + """Test RayJob with Kueue-managed lifecycled cluster with Secret validation.""" + self.setup_method() + create_namespace(self) + create_kueue_resources(self) + + self.job_api = RayjobApi() + cluster_api = RayClusterApi() + job_name = "kueue-lifecycled" + + # Get platform-appropriate resource configurations + resources = get_platform_appropriate_resources() + + cluster_config = ManagedClusterConfig( + head_cpu_requests=resources["head_cpu_requests"], + head_cpu_limits=resources["head_cpu_limits"], + head_memory_requests=resources["head_memory_requests"], + head_memory_limits=resources["head_memory_limits"], + num_workers=1, + worker_cpu_requests=resources["worker_cpu_requests"], + worker_cpu_limits=resources["worker_cpu_limits"], + worker_memory_requests=resources["worker_memory_requests"], + worker_memory_limits=resources["worker_memory_limits"], + ) + + # Create a temporary script file to test Secret functionality + with tempfile.NamedTemporaryFile( + mode="w", suffix=".py", delete=False, dir=os.getcwd() + ) as script_file: + script_file.write( + """ + import ray + ray.init() + print('Kueue job with Secret done') + ray.shutdown() + """ + ) + script_file.flush() + script_filename = os.path.basename(script_file.name) + + try: + rayjob = RayJob( + job_name=job_name, + namespace=self.namespace, + cluster_config=cluster_config, + entrypoint=f"python {script_filename}", + runtime_env={"env_vars": get_setup_env_variables(ACCELERATOR="cpu")}, + local_queue=self.local_queues[0], + ) + + assert rayjob.submit() == job_name + + # Verify Secret was created with owner reference + self.verify_secret_with_owner_reference(rayjob) + + assert self.job_api.wait_until_job_running( + name=rayjob.name, k8s_namespace=rayjob.namespace, timeout=600 + ) + + assert self.job_api.wait_until_job_finished( + name=rayjob.name, k8s_namespace=rayjob.namespace, timeout=300 + ) + finally: + try: + rayjob.delete() + except Exception: + pass # Job might already be deleted + verify_rayjob_cluster_cleanup(cluster_api, rayjob.name, rayjob.namespace) + # Clean up the temporary script file + if "script_filename" in locals(): + try: + os.remove(script_filename) + except: + pass + + def test_lifecycled_kueue_resource_queueing(self): + """Test Kueue resource queueing with lifecycled clusters.""" + self.setup_method() + create_namespace(self) + create_limited_kueue_resources(self) + + self.job_api = RayjobApi() + cluster_api = RayClusterApi() + + # Get platform-appropriate resource configurations + resources = get_platform_appropriate_resources() + + cluster_config = ManagedClusterConfig( + head_cpu_requests=resources["head_cpu_requests"], + head_cpu_limits=resources["head_cpu_limits"], + head_memory_requests=resources["head_memory_requests"], + head_memory_limits=resources["head_memory_limits"], + num_workers=0, + ) + + job1 = None + job2 = None + try: + job1 = RayJob( + job_name="holder", + namespace=self.namespace, + cluster_config=cluster_config, + entrypoint='python -c "import ray; import time; ray.init(); time.sleep(15)"', + runtime_env={"env_vars": get_setup_env_variables(ACCELERATOR="cpu")}, + local_queue=self.local_queues[0], + ) + assert job1.submit() == "holder" + assert self.job_api.wait_until_job_running( + name=job1.name, k8s_namespace=job1.namespace, timeout=60 + ) + + job2 = RayJob( + job_name="waiter", + namespace=self.namespace, + cluster_config=cluster_config, + entrypoint='python -c "import ray; ray.init()"', + runtime_env={"env_vars": get_setup_env_variables(ACCELERATOR="cpu")}, + local_queue=self.local_queues[0], + ) + assert job2.submit() == "waiter" + + # Wait for Kueue to process the job + sleep(5) + job2_cr = self.job_api.get_job(name=job2.name, k8s_namespace=job2.namespace) + + # For RayJobs with managed clusters, check if Kueue is holding resources + job2_status = job2_cr.get("status", {}) + ray_cluster_name = job2_status.get("rayClusterName", "") + + # If RayCluster is not created yet, it means Kueue is holding the job + if not ray_cluster_name: + # This is the expected behavior + job_is_queued = True + else: + # Check RayCluster resources - if all are 0, it's queued + ray_cluster_status = job2_status.get("rayClusterStatus", {}) + desired_cpu = ray_cluster_status.get("desiredCPU", "0") + desired_memory = ray_cluster_status.get("desiredMemory", "0") + + # Kueue creates the RayCluster but with 0 resources when queued + job_is_queued = desired_cpu == "0" and desired_memory == "0" + + assert job_is_queued, "Job2 should be queued by Kueue while Job1 is running" + + assert self.job_api.wait_until_job_finished( + name=job1.name, k8s_namespace=job1.namespace, timeout=60 + ) + + assert wait_for_kueue_admission( + self, self.job_api, job2.name, job2.namespace, timeout=30 + ) + + assert self.job_api.wait_until_job_finished( + name=job2.name, k8s_namespace=job2.namespace, timeout=60 + ) + finally: + for job in [job1, job2]: + if job: + try: + job.delete() + verify_rayjob_cluster_cleanup( + cluster_api, job.name, job.namespace + ) + except: + pass + + def verify_secret_with_owner_reference(self, rayjob: RayJob): + """Verify that the Secret was created with proper owner reference to the RayJob.""" + v1 = client.CoreV1Api() + secret_name = f"{rayjob.name}-files" + + try: + # Get the Secret + secret = v1.read_namespaced_secret( + name=secret_name, namespace=rayjob.namespace + ) + + # Verify Secret exists + assert secret is not None, f"Secret {secret_name} not found" + + # Verify it contains the script + assert secret.data is not None, "Secret has no data" + assert len(secret.data) > 0, "Secret data is empty" + + # Verify owner reference + assert ( + secret.metadata.owner_references is not None + ), "Secret has no owner references" + assert ( + len(secret.metadata.owner_references) > 0 + ), "Secret owner references list is empty" + + owner_ref = secret.metadata.owner_references[0] + assert ( + owner_ref.api_version == "ray.io/v1" + ), f"Wrong API version: {owner_ref.api_version}" + assert owner_ref.kind == "RayJob", f"Wrong kind: {owner_ref.kind}" + assert owner_ref.name == rayjob.name, f"Wrong owner name: {owner_ref.name}" + assert ( + owner_ref.controller is True + ), "Owner reference controller not set to true" + assert ( + owner_ref.block_owner_deletion is True + ), "Owner reference blockOwnerDeletion not set to true" + + # Verify labels + assert secret.metadata.labels.get("ray.io/job-name") == rayjob.name + assert ( + secret.metadata.labels.get("app.kubernetes.io/managed-by") + == "codeflare-sdk" + ) + assert ( + secret.metadata.labels.get("app.kubernetes.io/component") + == "rayjob-files" + ) + + print(f"✓ Secret {secret_name} verified with proper owner reference") + + except client.rest.ApiException as e: + if e.status == 404: + raise AssertionError(f"Secret {secret_name} not found") + else: + raise e diff --git a/tests/e2e/support.py b/tests/e2e/support.py index fe9261a2..85b3dd35 100644 --- a/tests/e2e/support.py +++ b/tests/e2e/support.py @@ -2,6 +2,7 @@ import random import string import subprocess +from time import sleep from codeflare_sdk import get_cluster from kubernetes import client, config from codeflare_sdk.common.kubernetes_cluster.kube_api_helpers import ( @@ -27,11 +28,82 @@ def get_ray_cluster(cluster_name, namespace): raise +def is_openshift(): + """Detect if running on OpenShift by checking for OpenShift-specific API resources.""" + try: + api = client.ApiClient() + discovery = client.ApisApi(api) + # Check for OpenShift-specific API group + groups = discovery.get_api_versions().groups + for group in groups: + if group.name == "image.openshift.io": + return True + return False + except Exception: + # If we can't determine, assume it's not OpenShift + return False + + def get_ray_image(): - return os.getenv( - "RAY_IMAGE", - get_ray_image_for_python_version(warn_on_unsupported=False), - ) + """ + Get appropriate Ray image based on platform (OpenShift vs Kind/vanilla K8s). + + The tests marked with @pytest.mark.openshift can run on both OpenShift and Kind clusters + with Kueue installed. This function automatically selects the appropriate image: + - OpenShift: Uses the CUDA runtime image (quay.io/modh/ray:...) + - Kind/K8s: Uses the standard Ray image (rayproject/ray:VERSION) + + You can override this behavior by setting the RAY_IMAGE environment variable. + """ + # Allow explicit override via environment variable + if "RAY_IMAGE" in os.environ: + return os.environ["RAY_IMAGE"] + + # Auto-detect platform and return appropriate image + if is_openshift(): + return get_ray_image_for_python_version() + else: + # Use standard Ray image for Kind/vanilla K8s + return f"rayproject/ray:{constants.RAY_VERSION}" + + +def get_platform_appropriate_resources(): + """ + Get appropriate resource configurations based on platform. + + OpenShift with MODH images requires more memory than Kind with standard Ray images. + + Returns: + dict: Resource configurations with keys: + - head_cpu_requests, head_cpu_limits + - head_memory_requests, head_memory_limits + - worker_cpu_requests, worker_cpu_limits + - worker_memory_requests, worker_memory_limits + """ + if is_openshift(): + # MODH runtime images require more memory + return { + "head_cpu_requests": "1", + "head_cpu_limits": "1.5", + "head_memory_requests": 7, + "head_memory_limits": 8, + "worker_cpu_requests": "1", + "worker_cpu_limits": "1", + "worker_memory_requests": 5, + "worker_memory_limits": 6, + } + else: + # Standard Ray images require less memory + return { + "head_cpu_requests": "1", + "head_cpu_limits": "1.5", + "head_memory_requests": 7, + "head_memory_limits": 8, + "worker_cpu_requests": "1", + "worker_cpu_limits": "1", + "worker_memory_requests": 2, + "worker_memory_limits": 3, + } def get_setup_env_variables(**kwargs): @@ -147,6 +219,17 @@ def run_oc_command(args): return None +def run_kubectl_command(args): + try: + result = subprocess.run( + ["kubectl"] + args, capture_output=True, text=True, check=True + ) + return result.stdout.strip() + except subprocess.CalledProcessError as e: + print(f"Error executing 'kubectl {' '.join(args)}': {e}") + return None + + def create_cluster_queue(self, cluster_queue, flavor): cluster_queue_json = { "apiVersion": "kueue.x-k8s.io/v1beta1", @@ -161,9 +244,9 @@ def create_cluster_queue(self, cluster_queue, flavor): { "name": flavor, "resources": [ - {"name": "cpu", "nominalQuota": 9}, - {"name": "memory", "nominalQuota": "36Gi"}, - {"name": "nvidia.com/gpu", "nominalQuota": 1}, + {"name": "cpu", "nominalQuota": 20}, + {"name": "memory", "nominalQuota": "80Gi"}, + {"name": "nvidia.com/gpu", "nominalQuota": 2}, ], }, ], @@ -301,7 +384,6 @@ def create_kueue_resources( def delete_kueue_resources(self): - # Delete if given cluster-queue exists for cq in self.cluster_queues: try: self.custom_api.delete_cluster_custom_object( @@ -409,3 +491,234 @@ def assert_get_cluster_and_jobsubmit( assert job_list[0].submission_id == submission_id cluster.down() + + +def wait_for_kueue_admission(self, job_api, job_name, namespace, timeout=120): + print(f"Waiting for Kueue admission of job '{job_name}'...") + elapsed_time = 0 + check_interval = 5 + + while elapsed_time < timeout: + try: + job_cr = job_api.get_job(name=job_name, k8s_namespace=namespace) + + # Check if the job is no longer suspended + is_suspended = job_cr.get("spec", {}).get("suspend", False) + + if not is_suspended: + print(f"✓ Job '{job_name}' admitted by Kueue (no longer suspended)") + return True + + # Debug: Check workload status every 10 seconds + if elapsed_time % 10 == 0: + workload = get_kueue_workload_for_job(self, job_name, namespace) + if workload: + conditions = workload.get("status", {}).get("conditions", []) + print(f" DEBUG: Workload conditions for '{job_name}':") + for condition in conditions: + print( + f" - {condition.get('type')}: {condition.get('status')} - {condition.get('reason', '')} - {condition.get('message', '')}" + ) + + # Optional: Check status conditions for more details + conditions = job_cr.get("status", {}).get("conditions", []) + for condition in conditions: + if ( + condition.get("type") == "Suspended" + and condition.get("status") == "False" + ): + print( + f"✓ Job '{job_name}' admitted by Kueue (Suspended=False condition)" + ) + return True + + except Exception as e: + print(f"Error checking job status: {e}") + + sleep(check_interval) + elapsed_time += check_interval + + print(f"✗ Timeout waiting for Kueue admission of job '{job_name}'") + return False + + +def create_limited_kueue_resources(self): + print("Creating limited Kueue resources for preemption testing...") + + # Create a resource flavor with default (no special labels/tolerations) + resource_flavor = f"limited-flavor-{random_choice()}" + create_resource_flavor( + self, resource_flavor, default=True, with_labels=False, with_tolerations=False + ) + self.resource_flavors = [resource_flavor] + + # Create a cluster queue with very limited resources + # Adjust quota based on platform - OpenShift needs more memory + if is_openshift(): + # MODH images need more memory, so higher quota but still limited to allow only 1 job + cpu_quota = 3 + memory_quota = "15Gi" # One job needs ~8Gi head, allow some buffer + else: + # Standard Ray images - one job needs ~8G head + 500m submitter + cpu_quota = 3 + memory_quota = "10Gi" # Enough for one job (8G head + submitter), but not two + + cluster_queue_name = f"limited-cq-{random_choice()}" + cluster_queue_json = { + "apiVersion": "kueue.x-k8s.io/v1beta1", + "kind": "ClusterQueue", + "metadata": {"name": cluster_queue_name}, + "spec": { + "namespaceSelector": {}, + "resourceGroups": [ + { + "coveredResources": ["cpu", "memory"], + "flavors": [ + { + "name": resource_flavor, + "resources": [ + { + "name": "cpu", + "nominalQuota": cpu_quota, + }, + { + "name": "memory", + "nominalQuota": memory_quota, + }, + ], + } + ], + } + ], + }, + } + + try: + self.custom_api.create_cluster_custom_object( + group="kueue.x-k8s.io", + plural="clusterqueues", + version="v1beta1", + body=cluster_queue_json, + ) + print(f"✓ Created limited ClusterQueue: {cluster_queue_name}") + except Exception as e: + print(f"Error creating limited ClusterQueue: {e}") + raise + + self.cluster_queues = [cluster_queue_name] + + # Create a local queue + local_queue_name = f"limited-lq-{random_choice()}" + create_local_queue(self, cluster_queue_name, local_queue_name, is_default=True) + self.local_queues = [local_queue_name] + + print("✓ Limited Kueue resources created successfully") + + +def get_kueue_workload_for_job(self, job_name, namespace): + try: + # List all workloads in the namespace + workloads = self.custom_api.list_namespaced_custom_object( + group="kueue.x-k8s.io", + version="v1beta1", + plural="workloads", + namespace=namespace, + ) + + # Find workload with matching RayJob owner reference + for workload in workloads.get("items", []): + owner_refs = workload.get("metadata", {}).get("ownerReferences", []) + + for owner_ref in owner_refs: + if ( + owner_ref.get("kind") == "RayJob" + and owner_ref.get("name") == job_name + ): + workload_name = workload.get("metadata", {}).get("name") + print( + f"✓ Found Kueue workload '{workload_name}' for RayJob '{job_name}'" + ) + return workload + + print(f"✗ No Kueue workload found for RayJob '{job_name}'") + return None + + except Exception as e: + print(f"Error getting Kueue workload for job '{job_name}': {e}") + return None + + +def wait_for_job_status( + job_api, rayjob_name: str, namespace: str, expected_status: str, timeout: int = 30 +) -> bool: + """ + Wait for a RayJob to reach a specific deployment status. + + Args: + job_api: RayjobApi instance + rayjob_name: Name of the RayJob + namespace: Namespace of the RayJob + expected_status: Expected jobDeploymentStatus value + timeout: Maximum time to wait in seconds + + Returns: + bool: True if status reached, False if timeout + """ + elapsed_time = 0 + check_interval = 2 + + while elapsed_time < timeout: + status = job_api.get_job_status(name=rayjob_name, k8s_namespace=namespace) + if status and status.get("jobDeploymentStatus") == expected_status: + return True + + sleep(check_interval) + elapsed_time += check_interval + + return False + + +def verify_rayjob_cluster_cleanup( + cluster_api, rayjob_name: str, namespace: str, timeout: int = 60 +): + """ + Verify that the RayCluster created by a RayJob has been cleaned up. + Handles KubeRay's automatic suffix addition to cluster names. + + Args: + cluster_api: RayClusterApi instance + rayjob_name: Name of the RayJob + namespace: Namespace to check + timeout: Maximum time to wait in seconds + + Raises: + TimeoutError: If cluster is not cleaned up within timeout + """ + elapsed_time = 0 + check_interval = 5 + + while elapsed_time < timeout: + # List all RayClusters in the namespace + clusters = cluster_api.list_ray_clusters( + k8s_namespace=namespace, async_req=False + ) + + # Check if any cluster exists that starts with our job name + found = False + for cluster in clusters.get("items", []): + cluster_name = cluster.get("metadata", {}).get("name", "") + # KubeRay creates clusters with pattern: {job_name}-raycluster-{suffix} + if cluster_name.startswith(f"{rayjob_name}-raycluster"): + found = True + break + + if not found: + # No cluster found, cleanup successful + return + + sleep(check_interval) + elapsed_time += check_interval + + raise TimeoutError( + f"RayCluster for job '{rayjob_name}' was not cleaned up within {timeout} seconds" + ) diff --git a/tests/test_cluster_yamls/appwrapper/unit-test-all-params.yaml b/tests/test_cluster_yamls/appwrapper/unit-test-all-params.yaml index aa097dd9..fe07e331 100644 --- a/tests/test_cluster_yamls/appwrapper/unit-test-all-params.yaml +++ b/tests/test_cluster_yamls/appwrapper/unit-test-all-params.yaml @@ -19,6 +19,7 @@ spec: controller-tools.k8s.io: '1.0' key1: value1 key2: value2 + ray.io/cluster: aw-all-params name: aw-all-params namespace: ns spec: diff --git a/tests/test_cluster_yamls/kueue/aw_kueue.yaml b/tests/test_cluster_yamls/kueue/aw_kueue.yaml index f5f16406..92e5078d 100644 --- a/tests/test_cluster_yamls/kueue/aw_kueue.yaml +++ b/tests/test_cluster_yamls/kueue/aw_kueue.yaml @@ -13,6 +13,7 @@ spec: metadata: labels: controller-tools.k8s.io: '1.0' + ray.io/cluster: unit-test-aw-kueue name: unit-test-aw-kueue namespace: ns spec: diff --git a/tests/test_cluster_yamls/kueue/ray_cluster_kueue.yaml b/tests/test_cluster_yamls/kueue/ray_cluster_kueue.yaml index d6db4f4a..04331aed 100644 --- a/tests/test_cluster_yamls/kueue/ray_cluster_kueue.yaml +++ b/tests/test_cluster_yamls/kueue/ray_cluster_kueue.yaml @@ -13,6 +13,7 @@ spec: metadata: labels: controller-tools.k8s.io: '1.0' + ray.io/cluster: unit-test-cluster-kueue name: unit-test-cluster-kueue namespace: ns spec: diff --git a/tests/test_cluster_yamls/ray/default-appwrapper.yaml b/tests/test_cluster_yamls/ray/default-appwrapper.yaml index 27828163..1041f3b5 100644 --- a/tests/test_cluster_yamls/ray/default-appwrapper.yaml +++ b/tests/test_cluster_yamls/ray/default-appwrapper.yaml @@ -11,6 +11,7 @@ spec: metadata: labels: controller-tools.k8s.io: '1.0' + ray.io/cluster: default-appwrapper name: default-appwrapper namespace: ns spec: diff --git a/tests/test_cluster_yamls/ray/default-ray-cluster.yaml b/tests/test_cluster_yamls/ray/default-ray-cluster.yaml index d2e6b05a..213b22cf 100644 --- a/tests/test_cluster_yamls/ray/default-ray-cluster.yaml +++ b/tests/test_cluster_yamls/ray/default-ray-cluster.yaml @@ -3,6 +3,7 @@ kind: RayCluster metadata: labels: controller-tools.k8s.io: '1.0' + ray.io/cluster: default-cluster name: default-cluster namespace: ns spec: diff --git a/tests/test_cluster_yamls/ray/unit-test-all-params.yaml b/tests/test_cluster_yamls/ray/unit-test-all-params.yaml index ee0878c1..7c7d82d6 100644 --- a/tests/test_cluster_yamls/ray/unit-test-all-params.yaml +++ b/tests/test_cluster_yamls/ray/unit-test-all-params.yaml @@ -10,6 +10,7 @@ metadata: key1: value1 key2: value2 kueue.x-k8s.io/queue-name: local-queue-default + ray.io/cluster: test-all-params name: test-all-params namespace: ns spec: