Rag Evaluation with Milvus, Ragas, Langchain, Ollama, Llama3

preview_player
Показать описание
About the session
Retrieval Augmented Generation (RAG) enhances chatbots by incorporating custom data in the prompt. Using large language models (LLMs) as judge has gained prominence in modern RAG systems. This talk will demo Ragas, an open-source automation tool for RAG evaluations. Christy will talk about and demo evaluating a RAG pipeline using Milvus and RAG metrics like context F1-score and answer correctness.

Topics Covered
- Foundation Model Evaluation vs RAG Evaluation
- Do you need human-labeled ground truths?
- Human Evaluation vs LLM-as-a-judge Evaluations
- Overall RAG vs RAG component Evaluations
- Example of different Retrieval methods with Evaluation
- Example of different Generation methods with Evaluation

▼ ▽ LINKS & RESOURCES

▼ ▽ JOIN THE COMMUNITY - MILVUS Discord channel

▶ CONNECT WITH US
Рекомендации по теме
Комментарии
Автор

Thank You for your informative seminar. I will also try to run your 'Demo Notebook'.

davidtindell
Автор

Great seminar, will run your notebook to learn more

gw