sequenceDiagram
participant Client
participant KuzcoWorker
participant VikeyAPI
participant Model/AI
Client->>KuzcoWorker: Request AI Service
KuzcoWorker->>VikeyAPI: Call Inference API
VikeyAPI->>Model/AI: Process Request
Model/AI->>VikeyAPI: Return Inference Result
VikeyAPI->>KuzcoWorker: Send API Response
KuzcoWorker->>Client: Deliver Final Result
https://github.com/arcxteam/kuzco-inference/blob/main/VIKEY-EN.md#flowchart-distribution
/kuzco-inference
├──.dockerignore
├──.ignore
│
├── /home
│ ├── .env
│ ├── Dockerfile
│ ├── docker-compose.yaml
│ ├── ex.nginx-conf
│ ├── execute.sh
│ └── start.sh
│
├── /vikey-inference
│ ├── .env
│ ├── Dockerfile
│ ├── docker-compose.yaml
│ ├── models.json
│ └── vikey-inference-linux # File config binary linux (for windows request)
│
-├── /dashboard # Realtime monitoring run (optional)
│ ├── Dockerfile
│ ├── docker-compose.yaml
│ ├── extract_log.py
│ ├── index.html # Homepage (check preview)
│ ├── monitor_server.py
│ └── inference_results.json
│
└── README.mdNote
Check preview homepage if need run (realtime) by own-localthost Kuzco Inference Dashboard
| Requirement | Details |
|---|---|
| Linux | Ubuntu 20 - 22 - 24 LTS |
| CPU | vCores 8 with 8GB RAM - more |
| STORAGE | Up 50GB - more spaces |
| API PROXY | Support LLM Ollama & OpenAI |
| PORT | Open 11434 (Official API Proxy) |
| Open 14444 (Official API Kuzco) | |
| Open 5050 (Custom for Dashboard) |
Obtain correct CLI WORKER
- Sign up here https://devnet.inference.net/register
- Navigate to create worker
- Create up to you
Worker Name→CLI→Create Worker - Click your
Worker Name→ clickInstanceandLaunch worker→ in section Step 2: Run an instance of this workerCOPY YOUR CODE
Note, use Vikey AI
- Need API-KEY w/ compatible format Ollama or OpenAI (LLM)
- I use
VikeyAIfor default Kuzco modelllama-3.2-3b-instruct& more models Rp.3000-5000/1M Tokens ~ $0.18-$0.3/1M Tokens - Readme for detail simple step bahasa indonesia
- Readme for detail simple step languange english
sudo apt update && sudo apt upgrade -y && \
sudo apt install screen curl ufw nload tree iptables git wget lz4 jq make gcc nano automake autoconf \
htop tmux libgbm1 protobuf-compiler python3 python3-pip python3-venv python3-dev python3-setuptools \
tar clang nethogs ncdu unzip build-essential pkg-config libssl-dev libleveldb-dev \
speedtest-cli ca-certificates libffi-dev libsqlite3-dev -ysource <(wget -qO- https://raw.githubusercontent.com/arcxteam/w-ai-wombo/main/nodejs.sh)
curl -sSL https://raw.githubusercontent.com/arcxteam/succinct-prover/refs/heads/main/docker.sh | bashgit clone https://github.com/arcxteam/kuzco-inference.gitsudo ufw allow 11434 && sudo ufw allow 14444 && sudo ufw allow 5050 && sudo ufw reload && sudo ufw status verboseAfter completed all step, final config for next step run
- Setup directory
kuzco-inference/home/(Main Running Kuzco) - Setup directory
kuzco-inference/vikey-inference/(API Proxy Model Run) - Setup directory
kuzco-inference/dashboard/(Optional run if need realtime monitor)
NOTE: RUN 1st
- First running
API Proxy LLModels - Replace this API KEY
VIKEY_API_KEY=vk-xxxxxxxxx - If success build check logs
docker compose logsis optional
cd kuzco-inference/vikey-inference && nano .envdocker compose up --build -dNOTE: RUN 2nd
- Seconds running
Main Run Kuzco official w/ config binary - After success run above, in terminal back to
cdand next step - Replace this
WORKER_CODE=xxxxxxxandWORKER_NAME=xxxxxxandVIKEY_API_KEY=vk-xxxxxx - Where do? worker code, worker name read above A.SETUP FOR ACCOUNT
- If success build check logs
docker compose logs -f
cd kuzco-inference/home && nano .envdocker compose up --build -dNOTE: AFTER SUCCESS ALL DIRECTORY RUN AND STABLE GET INFERENCE INFO LOGGING, YOU CAN RUN OWN DASHBOARD MONITORING THIS ACTUAL BY YOUR RUNNING LOG
- Get ip address
curl ifconfig.me && echo - For access web
http://ip-address:5050
cd kuzco-inference/dashboard && docker compose up --build -d# Check info logging
+ docker logs -f kuzco-inference --tail=200
+ docker compose logs -f --tail=200
# Stop & up, Go to each directory
+ cd kuzco-inference/home
+ cd kuzco-inference/dashboard
+ cd kuzco-inference/vikey-inference
+ docker compose down -v
+ docker compose up --build -d
# Delete total image build and clean
+ docker rmi home-kuzco-main:latest && docker rmi vikey-inference-vikey-inference:latest- Created here https://github.com/arcxteam/kuzco-inference/issues