Skip to content

Issues: vllm-project/vllm

[Roadmap] vLLM Roadmap Q1 2025
#11862 opened Jan 8, 2025 by simon-mo
Open 5
[V1] Feedback Thread
#12568 opened Jan 30, 2025 by simon-mo
Open 43
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

[Bug]: Error When VLLM_USE_TRITON_FLASH_ATTN=True on 2*8H100 bug Something isn't working
#13607 opened Feb 20, 2025 by phoenixsqf
1 task done
[Bug]: When using cpu inference, is the kv cache's physical memory space pre-allocated? bug Something isn't working
#13603 opened Feb 20, 2025 by 905799575
1 task done
[Installation]: how to use benchmarks in docker? installation Installation problems
#13598 opened Feb 20, 2025 by kkoren
1 task done
[Bug]: Marlin kernel doesn't work for multi-gpus bug Something isn't working
#13590 opened Feb 20, 2025 by meqiangxu
1 task done
[Bug]: arm64 No module named 'xformers' bug Something isn't working
#13585 opened Feb 20, 2025 by jiayi-1994
1 task done
[Feature]: Support for Running Classification Task in Online Server feature request good first issue Good for newcomers help wanted Extra attention is needed
#13567 opened Feb 19, 2025 by sam-h-bean
1 task done
[Bug]: Increasing root volume with guided decoding bug Something isn't working
#13556 opened Feb 19, 2025 by abpani
1 task done
[Usage]: How to use logits processors with max_num_seqs > 1? usage How to use vllm
#13553 opened Feb 19, 2025 by alejopaullier96
1 task done
[Bug]: there are some nccl erros when tp_size > 8 in offline inference bug Something isn't working
#13552 opened Feb 19, 2025 by yingtongxiong
1 task done
[Bug]: Make https://wheels.vllm.ai/nightly inspectable bug Something isn't working
#13545 opened Feb 19, 2025 by fxmarty-amd
1 task done
[Bug]: Ray+vllm run, then crash bug Something isn't working
#13535 opened Feb 19, 2025 by fantasy-mark
1 task done
[New Model]: facebook/contriever support requring help wanted Extra attention is needed new model Requests to new models
#13525 opened Feb 19, 2025 by yichuan520030910320
1 task done
[Bug]: Can't serve on ray cluster although passing VLLM_HOST_IP bug Something isn't working
#13521 opened Feb 19, 2025 by hahmad2008
1 task done
[Usage]: Does vllm support mix deploy on GPU+CPU? usage How to use vllm
#13517 opened Feb 19, 2025 by zengqingfu1442
1 task done
ProTip! no:milestone will show everything without a milestone.