A chatbot powered by AI and RAG (retrieval augmented generation). The chatbot takes into account docs specified by the user in order to provide more accurate answers.
The following instructions will show how to prepare the environment and run the chatbot locally. Also, it will show how to use the chatbot.
Download Ollama and install it in your machine. Pull the desidered LLM (llama2 will be used as an example).
ollama pull llama2
Create a directory called qdrant_storage then pull/run the qdrant container image with the following command:
podman run -p 6333:6333 -p 6334:6334 -e QDRANT_SERVICE_GRPC_PORT="6334" -v $(pwd)/qdrant_storage:/qdrant/storage:z docker://qdrant/qdrant
This section will show how to prepare and run the chatbot on the local environment.
Clone this repo with the following command:
git clone https://github.com/aguidirh/go-rag-chatbot.git
There are two configuration files. Both are under the configs folder in the root of go-rag-chat go mod.
config.yaml
apiVersion: v1alpha1
kind: GoRagChatbotConfig
spec:
qdrant:
host: http://0.0.0.0
port: 6333
collection: oc-mirror
llm:
model: llama2
scoreThreshold: 0.5
temperature: 0.8
server:
port: 8080
kb-config.yaml - contains an array of resources to be consumed as additional knowledge base for the llm (large language model)
apiVersion: v1alpha1
kind: GoRagChatbotKnowledgeBaseConfig
spec:
docs:
- https://docs.openshift.com/container-platform/4.16/installing/disconnected_install/about-installing-oc-mirror-v2.html
- https://github.com/openshift/oc-mirror/blob/main/v2/docs/enclave_support.md
- /assets/kb-docs
- https://github.com/openshift/oc-mirror/tree/main/v2/docs
The chatbot is exposed as an http server. Run the following command to bring it up:
go mod tidy
go run cmd/main.go
Additionally, podman-compose
or docker-compose
may be used to build and start go-rag-chatbot
. To do this, simply run:
mkdir qdrant_storage
podman-compose build
podman-compose up
Configuration will be derived from ./config
. docker-compose.yml
preferrentially sets up some arguments to ensure qdrant can be reached. Check
to see if the http endpoint is up by running:
$ curl 127.0.0.1:8080/healthz
ok
Check if qdrant is up by running:
$ curl 127.0.0.1:6333
{"title":"qdrant - vector search engine","version":"1.13.5","commit":"e282ed91e1f80a27cfa9d5d3d65b13b065b0eef8"
If saving time and network bandwidth when building the container image is important, it is possible to setup a GO Proxy. It will cache the dependencies and the podman-compose build will download them from there instead of getting them from the internet.
IMPORTANT: Run the steps below before podman-compose build.
In the example below it is used Athens, which is an open source enterprise ready Go Module proxy. Feel free to use other GO Proxys.
Step 1:
Create the directory where Athens will save the dependencies and export the path to the environment variable ATHENS_STORAGE.
export ATHENS_STORAGE=$HOME/athens-storage
mkdir -p $ATHENS_STORAGE
Step 2:
Run the Athens in a container with the following command:
podman run -d -v $ATHENS_STORAGE:/var/lib/athens:Z \
-e ATHENS_DISK_STORAGE_ROOT=/var/lib/athens \
-e ATHENS_STORAGE_TYPE=disk \
--name athens-goproxy \
--restart always \
-p 3000:3000 \
gomods/athens:v0.15.4
Step 3:
Export the GOPROXY environment variable, which will be used during the podman-compose build. The below config add http://0.0.0.0:3000
as the first option to get the dependency.
export GOPROXY=http://0.0.0.0:3000,proxy.golang.org,direct
This section will show how to load the vector database with docs and how to interect with the chatbot
A collection is where the documents are going to be stored in the vector database. It is necessary to create a collection before adding documents to the database.
Since the chatbot is exposed as an http server, it is possible to create a new collection calling the following URL:
/create-collection?collection-name=test-collection
Here is an example of a call using cURL:
curl "http://localhost:8080/create-collection?collection-name=alex-test"
The chatbot does not know about specific things related to your business/domain. It is possible to add domain specific document calling the following URL:
/add-docs
Here is an example of a call using cURL:
curl "http://localhost:8080/add-docs"
After adding the domain documents to the vector database, it is possible to ask questions to the chatbot about specific domain subjects. With the help of the documents added in the previous step, it is more than likely that the chatbot will know how to answer the question.
The chatbot is exposed in the following URL:
/chat
Here is an example of a call using cURL:
curl "http://localhost:8080/chat"