Skip to content

Pull requests: vllm-project/vllm

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

[Misc] Upgrade transformers to 4.49 ci/build documentation Improvements or additions to documentation
#13602 opened Feb 20, 2025 by ywang96 Loading…
Bump up the transformers version to v4.49.0 ci/build documentation Improvements or additions to documentation ready ONLY add when PR is ready to merge/full CI is needed
#13592 opened Feb 20, 2025 by WoosukKwon Loading…
fix neuron performance issue
#13589 opened Feb 20, 2025 by ajayvohra2005 Loading…
Integrating torchao quantization into vllm
#13588 opened Feb 20, 2025 by jerryzh168 Loading…
[Bugfix][CPU] Fix cpu all-reduce using native pytorch implementation ready ONLY add when PR is ready to merge/full CI is needed
#13586 opened Feb 20, 2025 by Isotr0py Loading…
[NVIDIA] Support nvfp4 cutlass gemm ci/build
#13571 opened Feb 19, 2025 by kaixih Loading…
[WIP] Use uv python for docker ci/build
#13569 opened Feb 19, 2025 by mgoin Loading…
[Bugfix] Fix quantization skip modules logic
#13562 opened Feb 19, 2025 by jeejeelee Draft
2 tasks
Remove unused kwargs from model definitions documentation Improvements or additions to documentation ready ONLY add when PR is ready to merge/full CI is needed speculative-decoding v1
#13555 opened Feb 19, 2025 by hmellor Loading…
[XPU]fix setuptools version for xpu ci/build
#13548 opened Feb 19, 2025 by yma11 Loading…
[Frontend] Add backend-specific options for guided decoding documentation Improvements or additions to documentation ready ONLY add when PR is ready to merge/full CI is needed structured-output
#13505 opened Feb 19, 2025 by joerunde Loading…
ProTip! Type g i on any issue or pull request to go back to the issue listing page.