Full demo and setup guide: https://youtu.be/Yl-T3YgePmw?si=n-vaZzClw0Q833E5
- Chatbot, Translation and Text-to-Speech, all completely free and running locally.
- Support voice output in Japanese, English, German, Spanish, French, Russian and more, powered by RVC, silero and voicevox.
- Includes custom finetuned model to avoid generic chatbot responses and breaking character.
- Gradio UI web interface.
- plugin support for easily adding other providers.
Manual setup (Tutorial video: Full demo and setup guide)
install python 3.10 https://www.python.org/downloads/release/python-3100/
install CUDA toolkit 12.4 https://developer.nvidia.com/cuda-12-4-0-download-archive
install visual studio and add desktop development with C++ component https://visualstudio.microsoft.com/downloads/
1. Download the project from releases
python -m venv venv
.\venv\Scripts\activate
(If you encounter an error that says “cannot be loaded because the execution of scripts is disabled on this system. Open powershell with admin privilage and run Set-ExecutionPolicy RemoteSigned
)
pip install -r requirements.txt
pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu124
pip install llama-cpp-python
pip install nltk
python -m nltk.downloader -d C:\nltk_data all
python main.py
When you see this message, go to http://localhost:7860 to see web UI
```
Running on local URL: http://127.0.0.1:7860
To create a public link, set `share=True` in `launch()`.
```
To start the program again, run:
.\venv\Scripts\activate
python main.py
If you have a decent GPU, You can install the GPU version of llama-cpp-python:
$env:CMAKE_ARGS ="-DGGML_CUDA=ON"
pip install llama-cpp-python --force-reinstall --no-cache-dir --verbose --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/124
This can improve latency further.
Just a reminder for someone using Command Prompt instead of Power Shell Change this
$env:CMAKE_ARGS ="-DGGML_CUDA=ON"
To this
set CMAKE_ARGS=-DGGML_CUDA=ON
or it will show error
$env:CMAKE_ARGS ="-DGGML_CUDA=ON"
The filename, directory name, or volume label syntax is incorrect.
- Download the project from releases
- Extract and double click run.bat
- When you see this message, go to http://localhost:7860 to see web UI
Running on local URL: http://127.0.0.1:7860
To create a public link, set `share=True` in `launch()`.
- Fetch chat input from streaming platforms (Finished)
- Improve local LLM (Finetuned model avaliable https://huggingface.co/xiaoheiqaq/Aya-7b-gguf)
- Write plugins for cloud providers(Azure tts, elevenlabs, chatgpt, whisper...)
- GPU support (Finished)
- Vtube studio integration (Finished)
- Let AI play games and provide commentary. (can currently play chess and keep talking nobody explode)
- AI singing
- NameError: name '_in_projection' is not defined
You cannot enable gpt sovits and rvc at the same time, some of their modules have conflict.
-
UnboundLocalError: local variable 'response' referenced before assignment
If you cloned this repo, you maybe missing model files for gpt-sovits, which will be in the zip folder in the releases section. replace plugins\gpt_sovits\models with the one from the zip.
-
To fetch chat from Youtube, copy the youtube_video_id from the stream url like this:
Then press start fetching chat