This is a Conversational Retrieval Augmented Generation (RAG) Knowledge Base Chat built on top of LLAMA2 (Embeddings & Model), Langchain and ChromaDB and orchestrated by FastAPI framework to provide and Endpoint for easy communication.
- Setup a virtual environement & Install the requirements:
python -m venv .venv
source .venv/bin/activate
pip install -r requirements.txt`
-
Copy your PDF files to the
documents
folder. -
Run the FastAPI server, to process and ingest your data on start with the LLM RAG and return the answer:
python main.py "What is the invoice number value?"