-
-
Notifications
You must be signed in to change notification settings - Fork 7.8k
Issues: vllm-project/vllm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug]: Deepseek-R1 with DEP16 hangs after kv cache allocation
bug
Something isn't working
#19101
opened Jun 3, 2025 by
ptarasiewiczNV
1 task done
[Bug]: Unable to Run W4A16 GPTQ Quantized Models
bug
Something isn't working
#19098
opened Jun 3, 2025 by
mchambrec
1 task done
[RFC]: Response format extensions for structured outputs
RFC
structured-output
v1
#19097
opened Jun 3, 2025 by
aarnphm
1 task done
[Bug]: Internal Server Error: python3 openai_chat_completion_client_for_multimodal.py -c audio when using Qwen/Qwen2-Audio-7B-Instruct
bug
Something isn't working
#19083
opened Jun 3, 2025 by
IceForChoco
1 task done
[Bug]: ModuleNotFoundError: No module named 'pandas'
bug
Something isn't working
#19078
opened Jun 3, 2025 by
ZhangShuaiyi
1 task done
2
[Bug]: CUDA error: unknown error when running vllm serve on WSL2 Ubuntu22.04
bug
Something isn't working
#19077
opened Jun 3, 2025 by
ezioasche
1 task done
[Feature]: Metal support
feature request
New feature or request
#19073
opened Jun 3, 2025 by
otarkhan
1 task done
[Bug]: vllm.third_party.pynvml.NVMLError_InvalidArgument: Invalid Argument
bug
Something isn't working
#19071
opened Jun 3, 2025 by
tengdecheng
1 task done
[Bug]: ValueError: Attempted to assign 119 = 119 multimodal tokens to 120 placeholders
bug
Something isn't working
#19070
opened Jun 3, 2025 by
jojolee123
1 task done
[Bug]: Device selection broken in v0.9
bug
Something isn't working
#19069
opened Jun 3, 2025 by
derfred
1 task done
[Usage]: how could I use vllm docker image in platform with arm64 tech cpu and nvidia a600 gpu
usage
How to use vllm
#19065
opened Jun 3, 2025 by
panys-aircas
1 task done
[Bug]: 'FutureWrapper' object has no attribute 'sampled_token_ids' when using ray to perform pipeline parallelism
bug
Something isn't working
#19063
opened Jun 3, 2025 by
havever
1 task done
[Bug]: Hermes tool parser stream output error in Qwen3 case
bug
Something isn't working
#19056
opened Jun 3, 2025 by
LiuLi1998
1 task done
[Bug]: vllm 0.9 image gives me gibberish
bug
Something isn't working
rocm
Related to AMD ROCm
#19052
opened Jun 3, 2025 by
azjam78910
[Bug]: 400 response when using Qwen3 + reasoning + tool calling + tool_choice "required"
bug
Something isn't working
#19051
opened Jun 3, 2025 by
andysalerno
1 task done
[Bug]: Quantization method specified in the model config (fp8) does not match the quantization method specified in the Something isn't working
quantization
argument (gguf).
bug
#19050
opened Jun 3, 2025 by
Minami-su
1 task done
[Bug]: System Memory OOM after upgrading to v0.9.0.1
bug
Something isn't working
#19048
opened Jun 3, 2025 by
ly0koS
1 task done
[Usage]: ValueError: Initial test run failed - Please make sure benchmark arguments are correctly specified. Error: Not Found
usage
How to use vllm
#19047
opened Jun 3, 2025 by
Estella31
1 task done
[Usage]: TorchDispatchMode does not work for vllm
usage
How to use vllm
#19044
opened Jun 3, 2025 by
helunwencser
1 task done
[Bug]: vllm profiling result contains invalid utf-8 code
bug
Something isn't working
#19043
opened Jun 3, 2025 by
helunwencser
1 task done
[Usage]: how to load multy models in one vllm process?
usage
How to use vllm
#19042
opened Jun 3, 2025 by
541543262
1 task done
[RFC]: Prefill-only optimizations for PD disaggregation in vLLM
RFC
#19038
opened Jun 2, 2025 by
KuntaiDu
1 task done
[Bug]: 100% CPU usage when idle. While loop in Something isn't working
acquire_read
pegging the CPU.
bug
#19036
opened Jun 2, 2025 by
MathieuBordere
1 task done
[Bug]: Dual a6000 pros not working. Arch 120.
bug
Something isn't working
#19025
opened Jun 2, 2025 by
vladrad
1 task done
Previous Next
ProTip!
Follow long discussions with comments:>50.