Skip to content

This repository contains the codebase for a chatbot powered by the RAG (Retrieval-Augmented Generation) process and integrated with ChromaDB for efficient document storage and retrieval.

Notifications You must be signed in to change notification settings

medhasreenivasan/RAG-empowered-LLMs

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

17 Commits
 
 
 
 
 
 

Repository files navigation

RAG-empowered-LLMs

This repository contains the codebase for a chatbot powered by the RAG (Retrieval-Augmented Generation) process and integrated with ChromaDB for efficient document storage and retrieval.

Setting up the ChromaDB

If you wish to set up the chromadb server locally follow the below steps

  1. To start the chromadb server run the below command: chroma run --path <path to your DB>
  2. To add, delete or create a new collection use the manage_collection.py script python manage_collection.py NEW <collection_name> <file_path>

If you wish to use chromadb on colab/local with a persist directory follow the below steps:

  1. run the script RAG/manage_collection.py to add, delete or create a new collection
  2. Use a persistent client to access the data chromadb.PersistentClient(path=<path to DB>)

Running the chatbot application

The functionality currently uses mistralai/Mistral-7B-Instruct-v0.2 model from huggingface for the RAG application.

  1. Run the application python app.py to get the gradio interface

Evaluation

To perform evaluation run eval.py For evaluating the RAG pipeline, we have adapted the metrics from ragas repository. For more details on evaluation check out https://github.com/explodinggradients/ragas

About

This repository contains the codebase for a chatbot powered by the RAG (Retrieval-Augmented Generation) process and integrated with ChromaDB for efficient document storage and retrieval.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages