Advanced RAG with ColBERT in LangChain and LlamaIndex

preview_player
Показать описание
ColBERT is a fast and accurate retrieval model, enabling scalable BERT-based search over large text collections in tens of milliseconds. This can be used as a potential alternative to Dense Embeddings in Retrieval Augmented Generation. In this video we explore using ColBERTv2 with RAGatouille and compare it with OpenAI Embedding models.

Signup for Advanced RAG:

LINKS:

TIMESTAMPS:
[00:00] Introduction
[00:29] Use ColBERT in LangChain
[08:46] Use ColBERT in LlamaIndex

All Interesting Videos:

Рекомендации по теме
Комментарии
Автор

can you make a video on how to evaluate a RAG? And compare different RAG approaches.

pawan
Автор

Very interesting video indeed!! Could you please create a video on how to use colBERTv2 for embedding with pg_vector for persistent storage?

mjaym
Автор

I’d like to do RAG over a medical textbook. What strategies would you recommend for chunking. I’m thinking a hierarchical graph structure makes intuitive sense. What are your thoughts on this?

hamslammula
Автор

@engineerprompt, can we use a persistant vector db like chroma, qdrant and others with Ragatouille? So that I can just embed the documents once and re-use them for inferences later.

nbbhaskar
Автор

Is the Plaid DB persistent? As in, if I do this, how do I connect to that particular DB again?

jayethompson
Автор

Cant find the google collab notebook? Would love to copy this across to my own account and havd a play. Not sure if I'm overlooking it? I just see the github link?

iaincampbell
Автор

How can we use approaches like ColBERT with other languages, as portuguese? Thanks!

almirbolduan
Автор

I am working on a machine that is running Ubuntu and connected to 4 80GB A100 GPU's. The issue i face is RAG.index cell is running forever on this machine. Whereas same code running on Google Colab free version runs within seconds. Any insights on how this can be resolved will be helpful. Thanks :)

shameekm
Автор

Navigating the landscape of storytelling and video experimentation, VideoGPT silently empowers my creative journey, adding a layer of sophistication to my content.

hales
Автор

of course the last result is more accurate. you gave it almost 50% (5 instead of 3 chunks) more context. when using multiple ways to achieve the same goal, please use the same amount of data. otherwise it is hard to compare the output.

on the topic of chunks given to RAG - why define that? what if one does not know about how many parts may contain relevant information?

iham
Автор

How can I monetize whatever is being said as a beginner..?

linuswatiti