Skip to content

chore: upgrade vLLM to 0.16.0 from GitHub releases#1893

Open
samsja wants to merge 4 commits intomainfrom
chore/vllm-0.16.0
Open

chore: upgrade vLLM to 0.16.0 from GitHub releases#1893
samsja wants to merge 4 commits intomainfrom
chore/vllm-0.16.0

Conversation

@samsja
Copy link
Member

@samsja samsja commented Feb 26, 2026

Summary

  • Upgrade vLLM from 0.16.0rc3 (S3 wheel) to 0.16.0 official release (GitHub)
  • Update flash-attn to torch 2.9 compatible build (2.8.3+cu129torch2.9)

Changes

Package Before After
vLLM 0.16.0rc3 0.16.0
flash-attn 2.8.3+cu128torch2.10 2.8.3+cu129torch2.9
torch 2.10.0 2.9.1

Test plan

  • Run inference server
  • Run training

🤖 Generated with Claude Code


Note

Medium Risk
Upgrades core GPU/inference dependencies (vllm, flash-attn, and the resolved torch/triton stack), which can change runtime behavior and binary compatibility for training/inference despite no application code changes.

Overview
Moves vllm from a pinned pre-release S3 wheel (0.16.0rc3) to the official PyPI release (vllm>=0.16.0) and removes the custom tool.uv.sources override for it.

Updates the flash-attn optional dependency to a Torch 2.9/CUDA 12.9-compatible wheel, and refreshes uv.lock accordingly (notably resolving to torch/torchaudio/torchvision 2.9.1 and triton 3.5.1, plus related NVIDIA package version shifts).

Written by Cursor Bugbot for commit 73e6873. This will update automatically on new commits. Configure here.

- Update vLLM from 0.16.0rc3 (S3) to 0.16.0 (GitHub release)
- Update flash-attn to torch 2.9 compatible build (cu129)

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
vLLM 0.16.0 is now available on PyPI, so we can use the official
package instead of the GitHub release wheel.

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
Co-authored-by: rasdani <73563550+rasdani@users.noreply.github.com>
Signed-off-by: samsja <55492238+samsja@users.noreply.github.com>
Copy link

@cursor cursor bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Cursor Bugbot has reviewed your changes and found 1 potential issue.

Bugbot Autofix is OFF. To automatically fix reported issues with cloud agents, enable autofix in the Cursor dashboard.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants