Skip to content

NebeyouMusie/End-To-End-Advanced-RAG-Project-using-Open-Source-LLM-Models-And-Groq-Inferencing

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

End To End Advanced RAG Project using Open Source LLM Models And Groq Inferencing

  • In this project I have built an end to end advanced RAG project using open source llm model, Mistral using groq inferencing engine.

Groq for RAG Image

DEMO

  • You can try the project live here

Description

  • This project showcase the implementation of an advanced RAG system that uses groq as an llm to retrieve information about langsmith.

Steps I followed:

  1. I have used the WebBaseLoader from the langchain_community document loader to load the data from the https://docs.smith.langchain.com/ webpage.
  2. transformed each text into a chunk of 1000 using the RecursiveCharacterTextSplitter imported from the langchain.text_splitter
  3. stored the vector embeddings which were made using the HuggingFaceInstructEmbeddings using the FAISS vector store.
  4. setup the llm ChatGroq with the model name mixtral-8x7b-32768
  5. Setup ChatPromptTemplate
  6. finally created the document_chain and retrieval_chain for chaining llm to prompt and retriever to document_chain respectively

Libraries Used

  • langchain==0.1.20
  • langchain-community==0.0.38
  • langchain-core==0.1.52
  • langchain-groq==0.1.3
  • faiss-cpu==1.8.0
  • python-dotenv

Installation

  1. Prerequisites
    • Git
    • Command line familiarity
  2. Clone the Repository: git clone https://github.com/NebeyouMusie/End-To-End-Advanced-RAG-Project-using-Open-Source-LLM-Models-And-Groq-Inferencing.git
  3. Create and Activate Virtual Environment (Recommended)
    • python -m venv venv
    • source venv/bin/activate
  4. Navigate to the projects directory cd ./End-To-End-Advanced-RAG-Project-using-Open-Source-LLM-Models-And-Groq-Inferencing using your terminal
  5. Install Libraries: pip install -r requirements.txt
  6. run streamlit run app.py
  7. open the link displayed in the terminal on your preferred browser

Collaboration

  • Collaborations are welcomed ❤️

Acknowledgments

Contact

About

In this project I have built an end to end advanced RAG project using open source llm model, Mistral using groq inferencing engine.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •  

Languages