Skip to content

waters222/exllamav2

Error
Looks like something went wrong!

About

A fast inference library for running LLMs locally on modern consumer-class GPUs

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 76.9%
  • Cuda 17.4%
  • C++ 5.7%