Skip to content

gyb357/langchain_chatbot

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

25 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

📑Introduction

Document-Based RAG-LangChain Chatbot

This repository implements a Retrieval-Augmented Generation (RAG) based chatbot system that runs entirely in a local environment. It processes various document formats such as PDF, TXT, and CSV to answer questions using the LangChain framework and provides an intuitive interface through Streamlit. The system is designed to work without relying on external services, making it suitable for private or sensitive data applications.

Purpose of the Project

This repository is focused on implementing LangChain and RAG in your local environment and testing simple techniques related to it.

Note

Testing will be done primarily on the dev branch, and if it's stable, it will be merged to the master branch.


✨Key Features

  • Process and extract data from various document formats (PDF, TXT, CSV)
  • Download and use embedding models and LLMs from Hugging Face
  • Build efficient search indices through document chunking and vectorization
  • Implement RAG-based question answering system
  • Configure LLM prompt chains using LangChain
  • Interactive web interface with Streamlit

🔍RAG Description

What is RAG-LangChain?

RAG-LangChain is an advanced AI framework that combines Retrieval-Augmented Generation (RAG) with LangChain, enabling more accurate, context-aware, and reliable AI-powered applications. Unlike traditional language models that generate responses solely based on their pre-trained knowledge, RAG-LangChain dynamically retrieves relevant information from external sources—such as documents, databases, and APIs—before generating answers. This approach significantly improves accuracy, reduces hallucinations, and enhances the explainability of responses.

How RAG Works

rag_image_0

  1. First, documents are processed, chunked, and converted into vector embeddings using embedding models. These vectors capture the semantic meaning of text.
  2. When a user asks a question, the system converts the question into the same vector space and retrieves the most relevant document chunks based on semantic similarity.

rag_image_1

  1. The retrieved context is then "augmented" to the user's query as additional context for the language model.
  2. Finally, the large language model (LLM) generates a response based on both the user's question and the retrieved document context.

(images from https://python.langchain.com/docs/tutorials/rag/)


🔨Getting Started

1. Check the Running Environment

  • 16GB ~ 32GB RAM recommended (32GB+ for larger models)
  • GPU acceleration (optional but recommended for performance)
  • Python 3.11+

Note

Based on relatively lightweight models ranging in size from 1B to 5B. Larger models will require more RAM and a high-performance GPU.

Important

You must ensure that the GPU is properly assigned as a processor.

2. Libraries and Dependencies

git clone https://github.com/gyb357/langchain_chatbot
pip install -r requirements.txt

3. Download LLMs and Managing Configuration

Access Huggingface to download models and browse for the right model for your task.

🤗 Huggingface: https://huggingface.co/models

This RAG system is local, and the embedding model and the large-scale language model (LLM) are managed through config.yaml. Instead of manually specifying the models in your code, you can define their names in the configuration file.

For example, within config/config.yaml, specify the embedding model and LLM model as follows:

embedding_model: "sentence-transformers/all-MiniLM-L6-v2"
llm_model: "mistralai/Mistral-7B-Instruct-v0.1"

4. Launching the Web Interface with Streamlit

To run streamlit, enter the command below to run it:

streamlit run app.py

After running this command, the application will be available at http://localhost:8501 by default. You can access the interface through your web browser.

If you're trying a different port number, follow the instructions below:

streamlit run [app.py] [--server.port 'Port number you want']

📁Project Structure

langchain_chatbot
├── app.py                      # Streamlit web application (main)
├── vectorstore/
│   └── vector_db.py            # Vector store related code
├── config/
│   └── config.yaml             # Configuration file
├── requirements.txt            # Dependency package list
└── README.md                   # Project documentation

About

Document-based RAG-LangChain Chatbot

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published