Skip to content

AYMNrh/RAG

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

All CPU, GPU-less RAG Model with Qdrant, Langchain, and GPT4All x Mistral-7B

This documentation outlines the integration of a Retriever-Augmented Generation (RAG) model that leverages the power of Qdrant for similarity search, Langchain for text splitting, and the GPT4All x Mistral-7B model for generating responses. This setup is designed to work entirely on CPU, offering a cost-effective solution for deploying advanced AI models.

Resources

Libraries Utilized

  • Langchain: Used for text splitting.
  • Qdrant: Utilized for similarity search and vector storage.
  • GPT4All: The backbone model platform.

Use Case: Recipe Recommendation System

This setup is demonstrated through a use case of a recipe recommendation system, showing the model's capability in handling and generating culinary recipes upon request.

Example Query

Entering new LLMChain chain...

Prompt: You are a chef expert bot. Below presents a context from which a question will be asked, give your valuable insights as well. Start by giving the name of the recipe, then the ingredients, then the recipe steps, always. And always end the recipe, don't stop before writing all steps.

Context: Oreo Birthday Cheesecake recipe details...

Question: Give me a recipe for a chocolate cake.

Answer:

Chocolate Cake Recipe Ingredients:

  • 2 cups all-purpose flour
  • 3/4 cup unsweetened cocoa powder ...

Instructions:

  1. Preheat oven to 350 degrees Fahrenheit...

Finished chain.

This documentation provides a comprehensive overview of the GPU-less RAG model integration using Qdrant, Langchain, and GPT4All x Mistral-7B, showcasing its application in a practical use case.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published