Skip to content

Issues: vllm-project/vllm

[Roadmap] vLLM Roadmap Q1 2025
#11862 opened Jan 8, 2025 by simon-mo
Open 5
[V1] Feedback Thread
#12568 opened Jan 30, 2025 by simon-mo
Open 47
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

[Usage]: Speculative Decoding KV Cache Generate usage How to use vllm
#13845 opened Feb 25, 2025 by qingyiaaaaa
1 task done
[Feature]: Add CLI Commands for Benchmarking feature request good first issue Good for newcomers
#13840 opened Feb 25, 2025 by simon-mo
1 task done
[Bug]: database disk image is malformed bug Something isn't working
#13838 opened Feb 25, 2025 by abpani
1 task done
[Feature]: Support triton 3.3.0 feature request
#13829 opened Feb 25, 2025 by johnnynunez
1 task done
[Bug]: Gibberish Output from LLaMA 3.1 8B using vLLM with xGrammar bug Something isn't working
#13828 opened Feb 25, 2025 by UTSAV-44
1 task done
[Bug]: vllm0.7.3: an illegal memory access was encountered bug Something isn't working
#13824 opened Feb 25, 2025 by tingjun-cs
1 task done
[Usage]: MLA disable prefix cache usage How to use vllm
#13822 opened Feb 25, 2025 by zeroorhero
[Bug]: Structured generation with JSON schema does not produce empty array bug Something isn't working
#13821 opened Feb 25, 2025 by Nithin-Holla
1 task done
[Performance]: Regarding PD separation performance performance Performance-related issues
#13816 opened Feb 25, 2025 by skyCreateXian
1 task done
[Bug]: vLLM serve Deepseek-R1-14B tensor parallelism failed bug Something isn't working
#13814 opened Feb 25, 2025 by caolonghao
1 task done
[Bug]: vLLM serve Deepseek-R1 on 4x8*L20 cluster failed bug Something isn't working
#13812 opened Feb 25, 2025 by ly01325
1 task done
[Usage]: vllm v0.7.2 can not support baichuan2 model usage How to use vllm
#13810 opened Feb 25, 2025 by alanayu
1 task done
[Feature]: Support DeepEP feature request
#13804 opened Feb 25, 2025 by fan-niu
1 task done
[Bug]: The accuracy of multiple cards and single card is inconsistent bug Something isn't working
#13801 opened Feb 25, 2025 by jiangix-paper
1 task done
[Bug]: ValueError: Unsupported FA version: None on V100 and V1 engine bug Something isn't working
#13788 opened Feb 24, 2025 by piotr-sikora-v
1 task done
[Bug]: vllm server hang when running DeepSeek R1 bug Something isn't working
#13778 opened Feb 24, 2025 by yangsijia-serena
1 task done
ProTip! What’s not been updated in a month: updated:<2025-01-25.