Skip to content

spotlight301/LLama-Local

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

14 Commits
 
 
 
 
 
 

Repository files navigation

LLAMA-Local-7b

LLAMA Local is a simple and interactive Streamlit application designed to generate AI responses. It uses the Python bindings for llama.cpp model for language understanding and generation. Users input a prompt, and the application returns a response generated by the model.

Getting Started

Dependencies This application requires the following Python libraries:

Streamlit Langchain You also need a Python 3.7+ environment.

Installation

Clone the repo:

bash

git clone https://github.com/Druvith/LLAMA-Local-An-AI-response-generator.git

Install the necessary packages:

pip install streamlit langchain
pip install llama-cpp-python

Download the GGML version of Llama-7B from huggingface.co and place it in the models directory, which should be in the same directory as your main project files.

Navigate to the project directory and run the Streamlit application:

cd PROJECT_NAME
streamlit run app2.py

Usage

After running the application, navigate to the local server address provided in your terminal (usually http://localhost:8501). In the app, you will find a text box where you can enter a prompt. After you submit the prompt, the application will display the AI's response.

Model Information

The application uses the python bindings for LlamaCpp language model, which is loaded from a local binary file (./models/llama-2-7b.ggmlv3.q4_K_S.bin). Ensure that this file is in the correct location for the application to function.

Contributing

Pull requests are welcome. For major changes, please open an issue first to discuss what you would like to change.

License MIT

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages