-
Notifications
You must be signed in to change notification settings - Fork 93
Issues: vllm-project/production-stack
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
bug: Model not found when enable vllm api key
bug
Something isn't working
#150
opened Feb 18, 2025 by
JustinDuy
Fix Installation Steps in vLLM Production Stack Tutorials
documentation
Improvements or additions to documentation
#37
opened Jan 29, 2025 by
dxu104
feat: Offline batched inference based on OpenAI offline batching API
feature request
New feature or request
#47
opened Jan 31, 2025 by
gaocegege
Helm Chart Lacks Clear Support for Multi-Node vLLM Deployment
help wanted
Extra attention is needed
#50
opened Jan 31, 2025 by
shohamyamin
Create an Example Building Ingress for Router Service
documentation
Improvements or additions to documentation
#60
opened Feb 4, 2025 by
0xThresh
Why Hugging Face Token?
question
Further information is requested
#67
opened Feb 6, 2025 by
nitin302
feat: Allow remote backend configuration
feature request
New feature or request
#75
opened Feb 7, 2025 by
askulkarni2
feat: Distributed tracing for router
feature request
New feature or request
help wanted
Extra attention is needed
#77
opened Feb 7, 2025 by
gaocegege
Feat: Router observability (Current QPS, router-side queueing delay, etc)
feature request
New feature or request
#78
opened Feb 7, 2025 by
sitloboi2012
Discussion: Unifying versions for helm and router
question
Further information is requested
#80
opened Feb 7, 2025 by
gaocegege
[Roadmap] vLLM production stack roadmap for 2025 Q1
#26
opened Jan 27, 2025 by
ApostaC
10 of 17 tasks
bug: flaky test case Functionality test for helm chart / Multiple-Models
bug
Something isn't working
#152
opened Feb 19, 2025 by
gaocegege
Discussion - QPS routing when there are multiple router replicas
discussion
question
Further information is requested
#166
opened Feb 21, 2025 by
aishwaryaraimule21
feature: Terraform Quickstart Tutorials for Underlying Infrastructure
feature request
New feature or request
#167
opened Feb 21, 2025 by
0xThresh
feature: Terraform Quickstart Tutorials for Google GKE
feature request
New feature or request
#172
opened Feb 23, 2025 by
falconlee236
feature: unify naming of production-stack, vllm-stack and vllm-router
discussion
feature request
New feature or request
#178
opened Feb 25, 2025 by
bufferoverflow
feature: introduce pyproject.toml and use uv
feature request
New feature or request
good first issue
Good for newcomers
help wanted
Extra attention is needed
#184
opened Feb 25, 2025 by
bufferoverflow
feature: custom callback functionality in vllm-router
feature request
New feature or request
help wanted
Extra attention is needed
#186
opened Feb 26, 2025 by
pwuersch
bug: File Access Error with vllm using runai_streamer on OCP
bug
Something isn't working
#193
opened Feb 27, 2025 by
TamKez
feature: Support CRD based configuration
feature request
New feature or request
#204
opened Mar 1, 2025 by
rootfs
bug-docs: Remove Something isn't working
sudo
from kubectl
and helm
commands
bug
#248
opened Mar 8, 2025 by
surajssd
Previous Next
ProTip!
Updated in the last three days: updated:>2025-03-06.