Skip to content

Update PyTorch to 2.7.0 #16859

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 39 commits into from
Apr 30, 2025
Merged
Show file tree
Hide file tree
Changes from 33 commits
Commits
Show all changes
39 commits
Select commit Hold shift + click to select a range
3498754
Update to PyTorch 2.7.0
huydhn Apr 18, 2025
a131606
Merge branch 'main' into torch-2.7.0-final-rc
huydhn Apr 18, 2025
52a7c4b
Install xformers from source
huydhn Apr 18, 2025
c33698e
Wrong installation order
huydhn Apr 18, 2025
363303b
Use --index-strategy unsafe-best-match temporarily
huydhn Apr 18, 2025
174c2a1
More tweak
huydhn Apr 18, 2025
1be359a
Try CUDA 12.6 then
huydhn Apr 21, 2025
32fe6df
Revert "Try CUDA 12.6 then"
huydhn Apr 21, 2025
a0e5706
More tweak for CPU, XPU and ROCm
huydhn Apr 22, 2025
3b27235
Leave XPU for now
huydhn Apr 22, 2025
100b6a7
Merge branch 'main' into torch-2.7.0-final-rc
huydhn Apr 22, 2025
079c59b
Use 2.7.0 final RC for now till it's ready on pypi
huydhn Apr 22, 2025
99b7173
Fix a minor import bug from XPU
huydhn Apr 22, 2025
1f62dc1
Another --index-strategy unsafe-best-match
huydhn Apr 22, 2025
2b70bf5
Install xformers and flashinfer from source
huydhn Apr 23, 2025
6e473d8
Add --no-build-isolation
huydhn Apr 23, 2025
72f27d0
Install xformers on vllm-base
huydhn Apr 23, 2025
86a6954
Keep xformers in cuda.txt and set TORCH_CUDA_ARCH_LIST
huydhn Apr 23, 2025
71ca7fc
Build xformers separately because it depends on torch (to build?)
huydhn Apr 23, 2025
2b9ee67
Update to PyTorch 2.7.0
huydhn Apr 24, 2025
b8e868e
Merge branch 'main' into torch-2.7.0
huydhn Apr 24, 2025
97ef2f0
Use https://download.pytorch.org/whl/test/cu128
huydhn Apr 24, 2025
a2d82e5
Revert "[CI/Build] workaround for CI build failure (#17070)"
huydhn Apr 24, 2025
c89b53d
Attempt to fix vllm serve on CPU
huydhn Apr 24, 2025
916013c
Set the correct parameter in helm
huydhn Apr 24, 2025
92b9710
Build mamba-ssm from source for now
huydhn Apr 24, 2025
ad39de9
Need to build flashinfer v0.2.2.post1 instead of latest main
huydhn Apr 24, 2025
9eff00c
Also pin xformers and mamba-ssm
huydhn Apr 24, 2025
197646b
Pin xformers doesn't work?
huydhn Apr 25, 2025
7c61758
Fix some more tests
huydhn Apr 25, 2025
ce11a99
Merge branch 'main' into torch-2.7.0-final-rc
huydhn Apr 25, 2025
1a4cc8c
Test pipeline typo
huydhn Apr 25, 2025
06b877e
Fix xformers https://github.com/facebookresearch/xformers/issues/1229
huydhn Apr 26, 2025
ca30743
Set --extra-index-url base on CUDA_VERSION
huydhn Apr 27, 2025
2c8ab23
Miss another spot
huydhn Apr 27, 2025
b50e2a0
Merge branch 'main' into torch-2.7.0-final-rc
huydhn Apr 29, 2025
1b1a4fa
Use the freshly bake xformers 0.0.30 from pypi
huydhn Apr 29, 2025
e0599be
Use CUDA 12.6 as default for release and nightly wheels
huydhn Apr 26, 2025
f6704fc
Let's go with CUDA 12.8
huydhn Apr 29, 2025
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion .buildkite/test-pipeline.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -449,7 +449,7 @@ steps:
- tests/models/encoder_decoder/language
commands:
# Install causal-conv1d for plamo2 models here, as it is not compatible with pip-compile.
- pip install causal-conv1d
- pip install 'git+https://github.com/Dao-AILab/causal-conv1d@v1.5.0.post8'
- pytest -v -s models/decoder_only/language -m 'core_model or quant_model'
- pytest -v -s models/embedding/language -m core_model

Expand Down
4 changes: 2 additions & 2 deletions .github/workflows/lint-and-deploy.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -66,7 +66,7 @@ jobs:
export AWS_SECRET_ACCESS_KEY=minioadmin
sleep 30 && kubectl -n ns-vllm logs -f "$(kubectl -n ns-vllm get pods | awk '/deployment/ {print $1;exit}')" &
helm install --wait --wait-for-jobs --timeout 5m0s --debug --create-namespace --namespace=ns-vllm test-vllm examples/online_serving/chart-helm -f examples/online_serving/chart-helm/values.yaml --set secrets.s3endpoint=http://minio:9000 --set secrets.s3bucketname=testbucket --set secrets.s3accesskeyid=$AWS_ACCESS_KEY_ID --set secrets.s3accesskey=$AWS_SECRET_ACCESS_KEY --set resources.requests.cpu=1 --set resources.requests.memory=4Gi --set resources.limits.cpu=2 --set resources.limits.memory=5Gi --set image.env[0].name=VLLM_CPU_KVCACHE_SPACE --set image.env[1].name=VLLM_LOGGING_LEVEL --set-string image.env[0].value="1" --set-string image.env[1].value="DEBUG" --set-string extraInit.s3modelpath="opt-125m/" --set-string 'resources.limits.nvidia\.com/gpu=0' --set-string 'resources.requests.nvidia\.com/gpu=0' --set-string image.repository="vllm-cpu-env"

- name: curl test
run: |
kubectl -n ns-vllm port-forward service/test-vllm-service 8001:80 &
Expand All @@ -79,4 +79,4 @@ jobs:
"max_tokens": 7,
"temperature": 0
}'):$CODE"
echo "$CODE"
echo "$CODE"
2 changes: 1 addition & 1 deletion .pre-commit-config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -46,7 +46,7 @@ repos:
rev: 0.6.2
hooks:
- id: pip-compile
args: [requirements/test.in, -o, requirements/test.txt]
args: [requirements/test.in, -o, requirements/test.txt, --index-strategy, unsafe-best-match]
files: ^requirements/test\.(in|txt)$
- repo: local
hooks:
Expand Down
4 changes: 2 additions & 2 deletions CMakeLists.txt
Original file line number Diff line number Diff line change
Expand Up @@ -46,8 +46,8 @@ set(HIP_SUPPORTED_ARCHS "gfx906;gfx908;gfx90a;gfx942;gfx950;gfx1030;gfx1100;gfx1
# requirements.txt files and should be kept consistent. The ROCm torch
# versions are derived from docker/Dockerfile.rocm
#
set(TORCH_SUPPORTED_VERSION_CUDA "2.6.0")
set(TORCH_SUPPORTED_VERSION_ROCM "2.6.0")
set(TORCH_SUPPORTED_VERSION_CUDA "2.7.0")
set(TORCH_SUPPORTED_VERSION_ROCM "2.7.0")

#
# Try to find python package with an executable that exactly matches
Expand Down
48 changes: 38 additions & 10 deletions docker/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -5,11 +5,11 @@
# docs/source/contributing/dockerfile/dockerfile.md and
# docs/source/assets/contributing/dockerfile-stages-dependency.png

ARG CUDA_VERSION=12.4.1
ARG CUDA_VERSION=12.8.1
#################### BASE BUILD IMAGE ####################
# prepare basic build environment
FROM nvidia/cuda:${CUDA_VERSION}-devel-ubuntu20.04 AS base
ARG CUDA_VERSION=12.4.1
ARG CUDA_VERSION=12.8.1
ARG PYTHON_VERSION=3.12
ARG TARGETPLATFORM
ENV DEBIAN_FRONTEND=noninteractive
Expand All @@ -34,6 +34,7 @@ RUN --mount=type=cache,target=/root/.cache/uv \
# This timeout (in seconds) is necessary when installing some dependencies via uv since it's likely to time out
# Reference: https://github.com/astral-sh/uv/pull/1694
ENV UV_HTTP_TIMEOUT=500
ENV UV_INDEX_STRATEGY="unsafe-best-match"

# Upgrade to GCC 10 to avoid https://gcc.gnu.org/bugzilla/show_bug.cgi?id=92519
# as it was causing spam when compiling the CUTLASS kernels
Expand Down Expand Up @@ -77,6 +78,13 @@ ENV TORCH_CUDA_ARCH_LIST=${torch_cuda_arch_list}
# Override the arch list for flash-attn to reduce the binary size
ARG vllm_fa_cmake_gpu_arches='80-real;90-real'
ENV VLLM_FA_CMAKE_GPU_ARCHES=${vllm_fa_cmake_gpu_arches}

# TESTING: install xformers from source until it's updated to work
# with 2.7.0. Also pin it to the current main commit to avoid the issue
# https://github.com/facebookresearch/xformers/issues/1229 with CUDA 12.8
RUN --mount=type=cache,target=/root/.cache/uv \
MAX_JOBS=16 TORCH_CUDA_ARCH_LIST='7.5 8.0 8.6 8.9 9.0+PTX' \
uv pip install --system --no-build-isolation "git+https://github.com/facebookresearch/xformers@f2de641ef670510cadab099ce6954031f52f191c"
#################### BASE BUILD IMAGE ####################

#################### WHEEL BUILD IMAGE ####################
Expand All @@ -89,6 +97,7 @@ COPY requirements/build.txt requirements/build.txt
# This timeout (in seconds) is necessary when installing some dependencies via uv since it's likely to time out
# Reference: https://github.com/astral-sh/uv/pull/1694
ENV UV_HTTP_TIMEOUT=500
ENV UV_INDEX_STRATEGY="unsafe-best-match"

RUN --mount=type=cache,target=/root/.cache/uv \
uv pip install --system -r requirements/build.txt
Expand Down Expand Up @@ -158,13 +167,15 @@ FROM base as dev
# This timeout (in seconds) is necessary when installing some dependencies via uv since it's likely to time out
# Reference: https://github.com/astral-sh/uv/pull/1694
ENV UV_HTTP_TIMEOUT=500
ENV UV_INDEX_STRATEGY="unsafe-best-match"

# Workaround for #17068
RUN --mount=type=cache,target=/root/.cache/uv \
uv pip install --system --no-build-isolation "git+https://github.com/state-spaces/[email protected]"

COPY requirements/lint.txt requirements/lint.txt
COPY requirements/test.txt requirements/test.txt
COPY requirements/dev.txt requirements/dev.txt
# Workaround for #17068
RUN --mount=type=cache,target=/root/.cache/uv \
uv pip install --system mamba-ssm==2.2.4 --no-build-isolation
RUN --mount=type=cache,target=/root/.cache/uv \
uv pip install --system -r requirements/dev.txt
#################### DEV IMAGE ####################
Expand All @@ -173,7 +184,7 @@ RUN --mount=type=cache,target=/root/.cache/uv \
# image with vLLM installed
# TODO: Restore to base image after FlashInfer AOT wheel fixed
FROM nvidia/cuda:${CUDA_VERSION}-devel-ubuntu22.04 AS vllm-base
ARG CUDA_VERSION=12.4.1
ARG CUDA_VERSION=12.8.1
ARG PYTHON_VERSION=3.12
WORKDIR /vllm-workspace
ENV DEBIAN_FRONTEND=noninteractive
Expand Down Expand Up @@ -203,6 +214,7 @@ RUN --mount=type=cache,target=/root/.cache/uv \
# This timeout (in seconds) is necessary when installing some dependencies via uv since it's likely to time out
# Reference: https://github.com/astral-sh/uv/pull/1694
ENV UV_HTTP_TIMEOUT=500
ENV UV_INDEX_STRATEGY="unsafe-best-match"

# Workaround for https://github.com/openai/triton/issues/2507 and
# https://github.com/pytorch/pytorch/issues/107960 -- hopefully
Expand All @@ -223,7 +235,15 @@ RUN --mount=type=cache,target=/root/.cache/uv \
# Install vllm wheel first, so that torch etc will be installed.
RUN --mount=type=bind,from=build,src=/workspace/dist,target=/vllm-workspace/dist \
--mount=type=cache,target=/root/.cache/uv \
uv pip install --system dist/*.whl --verbose
uv pip install --system dist/*.whl --verbose \
--extra-index-url https://download.pytorch.org/whl/cu128

# TESTING: install xformers from source until it's updated to work
# with 2.7.0. Also pin it to the current main commit to avoid the issue
# https://github.com/facebookresearch/xformers/issues/1229 with CUDA 12.8
RUN --mount=type=cache,target=/root/.cache/uv \
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

If the intention is that the we are just going to wait till xformers releases a 12.8 compat whl though ignore above

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

IMO, I think we can just build the package from source for CI, then switch to the official xformers package once it's ready

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

How long does this take to compile xformers? If it's too long I don't want to slow down our CI time for this.

Copy link
Contributor Author

@huydhn huydhn Apr 26, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Let me dig out the number for this once the build finish. Without caching, it would be significant from what I see locally, but let's see what it takes to build once this is cached.

MAX_JOBS=16 TORCH_CUDA_ARCH_LIST='7.5 8.0 8.6 8.9 9.0+PTX' \
uv pip install --system --no-build-isolation "git+https://github.com/facebookresearch/xformers@f2de641ef670510cadab099ce6954031f52f191c"

# If we need to build FlashInfer wheel before its release:
# $ export FLASHINFER_ENABLE_AOT=1
Expand All @@ -240,12 +260,18 @@ RUN --mount=type=bind,from=build,src=/workspace/dist,target=/vllm-workspace/dist
RUN --mount=type=cache,target=/root/.cache/uv \
. /etc/environment && \
if [ "$TARGETPLATFORM" != "linux/arm64" ]; then \
uv pip install --system https://github.com/flashinfer-ai/flashinfer/releases/download/v0.2.1.post2/flashinfer_python-0.2.1.post2+cu124torch2.6-cp38-abi3-linux_x86_64.whl ; \
# TESTING: install FlashInfer from source to test 2.7.0 final RC
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I would imagine FI has new release by now.

FLASHINFER_ENABLE_AOT=1 TORCH_CUDA_ARCH_LIST='7.5 8.0 8.6 8.9 9.0+PTX' \
uv pip install --system --no-build-isolation "git+https://github.com/flashinfer-ai/[email protected]" ; \
fi
COPY examples examples
COPY benchmarks benchmarks
COPY ./vllm/collect_env.py .

RUN --mount=type=cache,target=/root/.cache/uv \
. /etc/environment && \
uv pip list

# Although we build Flashinfer with AOT mode, there's still
# some issues w.r.t. JIT compilation. Therefore we need to
# install build dependencies for JIT compilation.
Expand All @@ -266,11 +292,13 @@ ADD . /vllm-workspace/
# This timeout (in seconds) is necessary when installing some dependencies via uv since it's likely to time out
# Reference: https://github.com/astral-sh/uv/pull/1694
ENV UV_HTTP_TIMEOUT=500
ENV UV_INDEX_STRATEGY="unsafe-best-match"

# install development dependencies (for testing)
# Workaround for #17068
RUN --mount=type=cache,target=/root/.cache/uv \
uv pip install --system mamba-ssm==2.2.4 --no-build-isolation
uv pip install --system --no-build-isolation "git+https://github.com/state-spaces/[email protected]"

# install development dependencies (for testing)
RUN --mount=type=cache,target=/root/.cache/uv \
uv pip install --system -r requirements/dev.txt

Expand Down
2 changes: 1 addition & 1 deletion examples/online_serving/chart-helm/values.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ image:
# -- Image tag
tag: "latest"
# -- Container launch command
command: ["vllm", "serve", "/data/", "--served-model-name", "opt-125m", "--dtype", "bfloat16", "--host", "0.0.0.0", "--port", "8000"]
command: ["vllm", "serve", "/data/", "--served-model-name", "opt-125m", "--dtype", "float32", "--block-size", "16", "--host", "0.0.0.0", "--port", "8000"]
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

this need to be reverted i assume

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This is a curious issue that I have seen on CI where cpu build fails on that dtype. I look around and see a similar issue which suggests switching to float32 #11327 (comment). Let me try to reproduce it to get the full log.

Copy link
Contributor Author

@huydhn huydhn Apr 26, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Here is the full server log when serving vllm serve facebook/opt-125m on the CPU docker image https://paste.sh/rDYEmrB9#9UuSY3jL6pHBE4MVy_lUDLle. The change you see here is only to make it work on CI, let me cut an issue #17225


# -- Container port
containerPort: 8000
Expand Down
2 changes: 1 addition & 1 deletion pyproject.toml
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ requires = [
"packaging",
"setuptools>=61",
"setuptools-scm>=8.0",
"torch == 2.6.0",
"torch == 2.7.0",
"wheel",
"jinja2",
]
Expand Down
3 changes: 2 additions & 1 deletion requirements/build.txt
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@ ninja
packaging
setuptools>=61
setuptools-scm>=8
torch==2.6.0
--extra-index-url https://download.pytorch.org/whl/cu128
torch==2.7.0
wheel
jinja2>=3.1.6
11 changes: 6 additions & 5 deletions requirements/cpu.txt
Original file line number Diff line number Diff line change
Expand Up @@ -2,18 +2,19 @@
-r common.txt

# Dependencies for CPUs
torch==2.6.0+cpu; platform_machine == "x86_64"
torch==2.6.0; platform_system == "Darwin"
torch==2.6.0; platform_machine == "ppc64le" or platform_machine == "aarch64"
--extra-index-url https://download.pytorch.org/whl/cpu
torch==2.7.0+cpu; platform_machine == "x86_64"
torch==2.7.0; platform_system == "Darwin"
torch==2.7.0; platform_machine == "ppc64le" or platform_machine == "aarch64"
torch==2.7.0.dev20250304; platform_machine == "s390x"

# required for the image processor of minicpm-o-2_6, this must be updated alongside torch
torchaudio; platform_machine != "ppc64le" and platform_machine != "s390x"
torchaudio==2.6.0; platform_machine == "ppc64le"
torchaudio==2.7.0; platform_machine == "ppc64le"

# required for the image processor of phi3v, this must be updated alongside torch
torchvision; platform_machine != "ppc64le" and platform_machine != "s390x"
torchvision==0.21.0; platform_machine == "ppc64le"
torchvision==0.22.0; platform_machine == "ppc64le"
datasets # for benchmark scripts

# cpu cannot use triton 3.3.0
Expand Down
11 changes: 7 additions & 4 deletions requirements/cuda.txt
Original file line number Diff line number Diff line change
Expand Up @@ -6,8 +6,11 @@ numba == 0.61.2; python_version > '3.9'

# Dependencies for NVIDIA GPUs
ray[cgraph]>=2.43.0, !=2.44.* # Ray Compiled Graph, required for pipeline parallelism in V1.
torch==2.6.0
torchaudio==2.6.0
--extra-index-url https://download.pytorch.org/whl/cu128
torch==2.7.0
torchaudio==2.7.0
# These must be updated alongside torch
torchvision==0.21.0 # Required for phi3v processor. See https://github.com/pytorch/vision?tab=readme-ov-file#installation for corresponding version
xformers==0.0.29.post2; platform_system == 'Linux' and platform_machine == 'x86_64' # Requires PyTorch 2.6.0
torchvision==0.22.0 # Required for phi3v processor. See https://github.com/pytorch/vision?tab=readme-ov-file#installation for corresponding version
# TESTING: comment out xformers for testing 2.7.0. It will be built from sources instead
# Note that building xformers from source requires torch
# xformers==0.0.29.post3; platform_system == 'Linux' and platform_machine == 'x86_64' # Requires PyTorch >= 2.6
6 changes: 3 additions & 3 deletions requirements/rocm-build.txt
Original file line number Diff line number Diff line change
Expand Up @@ -2,9 +2,9 @@
-r common.txt

--extra-index-url https://download.pytorch.org/whl/rocm6.2.4
torch==2.6.0
torchvision==0.21.0
torchaudio==2.6.0
torch==2.7.0
torchvision==0.22.0
torchaudio==2.7.0

triton==3.2
cmake>=3.26,<4
Expand Down
7 changes: 4 additions & 3 deletions requirements/test.in
Original file line number Diff line number Diff line change
Expand Up @@ -23,9 +23,10 @@ sentence-transformers # required for embedding tests
soundfile # required for audio tests
jiwer # required for audio tests
timm # required for internvl test
torch==2.6.0
torchaudio==2.6.0
torchvision==0.21.0
--extra-index-url https://download.pytorch.org/whl/cu128
torch==2.7.0
torchaudio==2.7.0
torchvision==0.22.0
transformers_stream_generator # required for qwen-vl test
mamba_ssm # required for plamo2 test
matplotlib # required for qwen-vl test
Expand Down
44 changes: 24 additions & 20 deletions requirements/test.txt
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
# This file was autogenerated by uv via the following command:
# uv pip compile requirements/test.in -o requirements/test.txt
# uv pip compile requirements/test.in -o requirements/test.txt --index-strategy unsafe-best-match
absl-py==2.1.0
# via rouge-score
accelerate==1.0.1
Expand Down Expand Up @@ -274,7 +274,7 @@ mamba-ssm==2.2.4
# via -r requirements/test.in
markdown-it-py==3.0.0
# via rich
markupsafe==3.0.2
markupsafe==3.0.1
# via
# jinja2
# werkzeug
Expand Down Expand Up @@ -355,39 +355,42 @@ numpy==1.26.4
# transformers
# tritonclient
# vocos
nvidia-cublas-cu12==12.4.5.8
nvidia-cublas-cu12==12.6.4.1
# via
# nvidia-cudnn-cu12
# nvidia-cusolver-cu12
# torch
nvidia-cuda-cupti-cu12==12.4.127
nvidia-cuda-cupti-cu12==12.6.80
# via torch
nvidia-cuda-nvrtc-cu12==12.4.127
nvidia-cuda-nvrtc-cu12==12.6.77
# via torch
nvidia-cuda-runtime-cu12==12.4.127
nvidia-cuda-runtime-cu12==12.6.77
# via torch
nvidia-cudnn-cu12==9.1.0.70
nvidia-cudnn-cu12==9.5.1.17
# via torch
nvidia-cufft-cu12==11.2.1.3
nvidia-cufft-cu12==11.3.0.4
# via torch
nvidia-curand-cu12==10.3.5.147
nvidia-cufile-cu12==1.11.1.6
# via torch
nvidia-cusolver-cu12==11.6.1.9
nvidia-curand-cu12==10.3.7.77
# via torch
nvidia-cusparse-cu12==12.3.1.170
nvidia-cusolver-cu12==11.7.1.2
# via torch
nvidia-cusparse-cu12==12.5.4.2
# via
# nvidia-cusolver-cu12
# torch
nvidia-cusparselt-cu12==0.6.2
nvidia-cusparselt-cu12==0.6.3
# via torch
nvidia-nccl-cu12==2.21.5
nvidia-nccl-cu12==2.26.2
# via torch
nvidia-nvjitlink-cu12==12.4.127
nvidia-nvjitlink-cu12==12.6.85
# via
# nvidia-cufft-cu12
# nvidia-cusolver-cu12
# nvidia-cusparse-cu12
# torch
nvidia-nvtx-cu12==12.4.127
nvidia-nvtx-cu12==12.6.77
# via torch
opencv-python-headless==4.11.0.86
# via
Expand Down Expand Up @@ -634,6 +637,7 @@ setuptools==75.8.0
# mamba-ssm
# pytablewriter
# torch
# triton
shellingham==1.5.4
# via typer
six==1.16.0
Expand Down Expand Up @@ -664,7 +668,7 @@ starlette-testclient==0.4.1
# via schemathesis
statsmodels==0.14.4
# via genai-perf
sympy==1.13.1
sympy==1.13.3
# via
# einx
# torch
Expand Down Expand Up @@ -696,7 +700,7 @@ tomli==2.2.1
# via schemathesis
tomli-w==1.2.0
# via schemathesis
torch==2.6.0
torch==2.7.0
# via
# -r requirements/test.in
# accelerate
Expand All @@ -714,12 +718,12 @@ torch==2.6.0
# torchvision
# vector-quantize-pytorch
# vocos
torchaudio==2.6.0
torchaudio==2.7.0+cu128
# via
# -r requirements/test.in
# encodec
# vocos
torchvision==0.21.0
torchvision==0.22.0+cu128
# via
# -r requirements/test.in
# timm
Expand Down Expand Up @@ -748,7 +752,7 @@ transformers==4.51.1
# transformers-stream-generator
transformers-stream-generator==0.0.5
# via -r requirements/test.in
triton==3.2.0
triton==3.3.0
# via torch
tritonclient==2.51.0
# via
Expand Down
Loading