Working on LLM inference systems, KV cache compression, and kernel-level optimizations (TurboQuant).
Pinned Loading
-
-
llama-cpp-turboquant
llama-cpp-turboquant PublicForked from ggml-org/llama.cpp
LLM inference in C/C++
-
elm327_obd_for_mac
elm327_obd_for_mac PublicI'm crazy and trying to make a ForScan OBD reader work on my mac.
-
vllm-swift
vllm-swift PublicvLLM Metal plugin powered by mlx-swift — high-performance LLM inference on Apple Silicon
-
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.




