🎯
Focusing
- Chengdu, China
-
22:28
- 8h ahead
Pinned Loading
-
vllm
vllm PublicForked from vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
Python 6
-
bitsandbytes
bitsandbytes PublicForked from bitsandbytes-foundation/bitsandbytes
Accessible large language models via k-bit quantization for PyTorch.
Python
-
flashinfer
flashinfer PublicForked from flashinfer-ai/flashinfer
FlashInfer: Kernel Library for LLM Serving
Cuda
511 contributions in the last year
Day of Week | April Apr | May May | June Jun | July Jul | August Aug | September Sep | October Oct | November Nov | December Dec | January Jan | February Feb | March Mar | |||||||||||||||||||||||||||||||||||||||||
Sunday Sun | |||||||||||||||||||||||||||||||||||||||||||||||||||||
Monday Mon | |||||||||||||||||||||||||||||||||||||||||||||||||||||
Tuesday Tue | |||||||||||||||||||||||||||||||||||||||||||||||||||||
Wednesday Wed | |||||||||||||||||||||||||||||||||||||||||||||||||||||
Thursday Thu | |||||||||||||||||||||||||||||||||||||||||||||||||||||
Friday Fri | |||||||||||||||||||||||||||||||||||||||||||||||||||||
Saturday Sat |
Less
No contributions.
Low contributions.
Medium-low contributions.
Medium-high contributions.
High contributions.
More
Activity overview
Contributed to
vllm-project/vllm,
jeejeelee/punica_triton_kernel,
flashinfer-ai/flashinfer
and 9 other
repositories
Loading
Contribution activity
April 2025
Created 4 commits in 1 repository
Created a pull request in vllm-project/vllm that received 7 comments
[CI/Build] Clean up LoRA tests
To reduce the testing time for LoRA and ensure coverage of all LoRA features in testing, the following modifications were made: Remove long contex…
+27
−200
lines changed
•
7
comments
Opened 4 other pull requests in 1 repository
vllm-project/vllm
3
merged
1
open
-
[CI/Build] Fix CI LoRA failure
This contribution was made on Apr 8
-
[Core][LoRA] Add LoRA for EncoderDecoderModelRunner
This contribution was made on Apr 3
-
[CI/Build] Further clean up LoRA tests
This contribution was made on Apr 2
-
[Misc] V1 LoRA support CPU offload
This contribution was made on Apr 1
Reviewed 18 pull requests in 1 repository
vllm-project/vllm
18 pull requests
-
Add GLM-4-0414 support
This contribution was made on Apr 9
-
[Model] Add
SupportsMultiModal.get_language_model
interfaceThis contribution was made on Apr 9 -
[CI/Build] Fix CI LoRA failure
This contribution was made on Apr 8
-
[Bugfix] Remove triton do_bench fast_flush arg
This contribution was made on Apr 8
-
[Bugfix] Fix and reorganize broken GGUF tests and bump gguf version
This contribution was made on Apr 8
-
[Misc] Merge the logs of pp layers partitions
This contribution was made on Apr 8
-
[Bugfix] Re-enable support for
ChatGLMForConditionalGeneration
This contribution was made on Apr 7 -
[Misc] Update Mistral-3.1 example
This contribution was made on Apr 7
-
[Bugfix] LoRA : Fix the order in which the kernels process LoRAs
This contribution was made on Apr 6
-
Fix some capitalisations in generated examples doc titles
This contribution was made on Apr 5
-
[Misc] Auto detect bitsandbytes pre-quantized models
This contribution was made on Apr 4
-
[misc] improve error message for "Failed to infer device type"
This contribution was made on Apr 3
-
[V1][Bugfix] Fix typo in MoE TPU checking
This contribution was made on Apr 2
-
[Model] Add module name prefixes to gemma3
This contribution was made on Apr 1
-
Remove
python_only_dev.py
as it's been unsupported for >110 daysThis contribution was made on Apr 1 -
Remove
format.sh
as it's been unsupported >70 daysThis contribution was made on Apr 1 -
[Hardware][TPU][V1] Multi-LoRA implementation for the V1 TPU backend
This contribution was made on Apr 1
-
[Bugfix]: Fix is_embedding_layer condition in VocabParallelEmbedding
This contribution was made on Apr 1