Creating a Retrieval Augmented Generation (RAG) system with Ollama, Llama3 and Langchain | RAG

preview_player
Показать описание

This YouTube video tutorial explains how to create a Retrieval-Augmented Generation (RAG) question-answering system using the LangChain library. The presenter introduces RAG, which combines Large Language Models (LLMs) with contextual information from external sources to enhance text generation. The specific project demonstrated involves creating an RAG system based on Adobe Photoshop documentation, which consists of 1017 pages.

The process includes:

Loading the PDF documentation using LangChain's PyPDFLoader.
Splitting the text into chunks with the RecursiveCharacterTextSplitter for better semantic retention.
Embedding the text chunks using a model from the Massive Text Embedding Benchmark (MTEB) leaderboard.
Implementing a vector search database with Facebook's FAISS (Facebook AI Similarity Search) to efficiently handle and retrieve similar text vectors.
The video also covers installing necessary packages and provides code snippets throughout the tutorial. The complete notebook and code will be shared in the video description.

Follow me here:
Рекомендации по теме
Комментарии
Автор

Can you explain more the cell 27, the db.as_retriever? Does that mean you are passing to the LLM the entire db? How much tokens does that eat up? I'm trying to understand pricing since I am using your example for a POC at work! :) Thank you!!!

lesptitsoiseaux