forked from turboderp-org/exllamav2
-
Notifications
You must be signed in to change notification settings - Fork 0
A fast inference library for running LLMs locally on modern consumer-class GPUs
License
waters222/exllamav2
ErrorLooks like something went wrong!
About
A fast inference library for running LLMs locally on modern consumer-class GPUs
Resources
License
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published
Languages
- Python 76.9%
- Cuda 17.4%
- C++ 5.7%