Skip to content

aguidirh/go-rag-chatbot

Repository files navigation

Overview

A chatbot powered by AI and RAG (retrieval augmented generation). The chatbot takes into account docs specified by the user in order to provide more accurate answers.

WARN: Working In Progress - Do not use it yet

Table of contents

Getting Started

The following instructions will show how to prepare the environment and run the chatbot locally. Also, it will show how to use the chatbot.

Requirements

Ollama

Download Ollama and install it in your machine. Pull the desidered LLM (llama2 will be used as an example).

ollama pull llama2

Qdrant

Create a directory called qdrant_storage then pull/run the qdrant container image with the following command:

podman run -p 6333:6333 -p 6334:6334 -e QDRANT_SERVICE_GRPC_PORT="6334" -v $(pwd)/qdrant_storage:/qdrant/storage:z docker://qdrant/qdrant

Setting up and running the ChatBot App

This section will show how to prepare and run the chatbot on the local environment.

Step 1 - Clone this repo

Clone this repo with the following command:

git clone https://github.com/aguidirh/go-rag-chatbot.git

Step 2 - Change the config files according to the needs

There are two configuration files. Both are under the configs folder in the root of go-rag-chat go mod.

config.yaml

apiVersion: v1alpha1
kind: GoRagChatbotConfig
spec:
  qdrant:
    host: http://0.0.0.0
    port: 6333
    collection: oc-mirror
  llm:
    model: llama2
    scoreThreshold: 0.5
    temperature: 0.8
  server:
    port: 8080

kb-config.yaml - contains an array of resources to be consumed as additional knowledge base for the llm (large language model)

apiVersion: v1alpha1
kind: GoRagChatbotKnowledgeBaseConfig
spec:
  docs:
    - https://docs.openshift.com/container-platform/4.16/installing/disconnected_install/about-installing-oc-mirror-v2.html
    - https://github.com/openshift/oc-mirror/blob/main/v2/docs/enclave_support.md
    - /assets/kb-docs
    - https://github.com/openshift/oc-mirror/tree/main/v2/docs

Step 3 - Run the ChatBot App

The chatbot is exposed as an http server. Run the following command to bring it up:

go mod tidy
go run cmd/main.go

Running as a Container

Additionally, podman-compose or docker-compose may be used to build and start go-rag-chatbot. To do this, simply run:

mkdir qdrant_storage
podman-compose build
podman-compose up

Configuration will be derived from ./config. docker-compose.yml preferrentially sets up some arguments to ensure qdrant can be reached. Check to see if the http endpoint is up by running:

$ curl 127.0.0.1:8080/healthz
ok

Check if qdrant is up by running:

$ curl 127.0.0.1:6333
{"title":"qdrant - vector search engine","version":"1.13.5","commit":"e282ed91e1f80a27cfa9d5d3d65b13b065b0eef8"

Using a GO proxy (Optional)

If saving time and network bandwidth when building the container image is important, it is possible to setup a GO Proxy. It will cache the dependencies and the podman-compose build will download them from there instead of getting them from the internet.

IMPORTANT: Run the steps below before podman-compose build.

In the example below it is used Athens, which is an open source enterprise ready Go Module proxy. Feel free to use other GO Proxys.

Step 1:

Create the directory where Athens will save the dependencies and export the path to the environment variable ATHENS_STORAGE.

export ATHENS_STORAGE=$HOME/athens-storage
mkdir -p $ATHENS_STORAGE

Step 2:

Run the Athens in a container with the following command:

podman run -d -v $ATHENS_STORAGE:/var/lib/athens:Z    \
-e ATHENS_DISK_STORAGE_ROOT=/var/lib/athens    \
-e ATHENS_STORAGE_TYPE=disk       \
--name athens-goproxy    \
--restart always    \
-p 3000:3000    \
gomods/athens:v0.15.4

Step 3:

Export the GOPROXY environment variable, which will be used during the podman-compose build. The below config add http://0.0.0.0:3000 as the first option to get the dependency.

export GOPROXY=http://0.0.0.0:3000,proxy.golang.org,direct

Using the Chatbot

This section will show how to load the vector database with docs and how to interect with the chatbot

Creating a collection

A collection is where the documents are going to be stored in the vector database. It is necessary to create a collection before adding documents to the database.

Since the chatbot is exposed as an http server, it is possible to create a new collection calling the following URL:

/create-collection?collection-name=test-collection

Here is an example of a call using cURL:

curl "http://localhost:8080/create-collection?collection-name=alex-test"

Adding documents to the vector database

The chatbot does not know about specific things related to your business/domain. It is possible to add domain specific document calling the following URL:

/add-docs

Here is an example of a call using cURL:

curl "http://localhost:8080/add-docs"

Chating with the chat bot

After adding the domain documents to the vector database, it is possible to ask questions to the chatbot about specific domain subjects. With the help of the documents added in the previous step, it is more than likely that the chatbot will know how to answer the question.

The chatbot is exposed in the following URL:

/chat

Here is an example of a call using cURL:

curl "http://localhost:8080/chat"

Built With

Author