Running Llama 2 and other Open-Source LLMs on CPU Inference Locally for Document Q&A
-
Updated
Nov 6, 2023 - Python
Running Llama 2 and other Open-Source LLMs on CPU Inference Locally for Document Q&A
Runs LLaMA with Extremely HIGH speed
LLM inference in Fortran
The bare metal in my basement
Portable LLM - A rust library for LLM inference
Wrapper for simplified use of Llama2 GGUF quantized models.
Simple large language model playground app
Run Mistral, LLaMA, and DeepSeek locally on Windows with zero setup — no Python required.
V-lang api wrapper for llm-inference chatllm.cpp
VB.NET api wrapper for llm-inference chatllm.cpp
C# api wrapper for llm-inference chatllm.cpp
Nim api-wrapper for llm-inference chatllm.cpp
eLLM provides million-token inference on CPUs
Simple bot that transcribes Telegram voice messages. Powered by go-telegram-bot-api & whisper.cpp Go bindings.
🧠 A comprehensive toolkit for benchmarking, optimizing, and deploying local Large Language Models. Includes performance testing tools, optimized configurations for CPU/GPU/hybrid setups, and detailed guides to maximize LLM performance on your hardware.
rust api wrapper for llm-inference chatllm.cpp
lua api wrapper for llm-inference chatllm.cpp
kotlin api wrapper for llm-inference chatllm.cpp
gemma-2-2b-it int8 cpu inference in one file of pure C#
Java-port of qwen3.c
Add a description, image, and links to the cpu-inference topic page so that developers can more easily learn about it.
To associate your repository with the cpu-inference topic, visit your repo's landing page and select "manage topics."