Evaluate LLM Systems & RAGs: Choose the Best LLM Using Automatic Metrics on Your Dataset

preview_player
Показать описание
Learn how to effectively evaluate new Large Language Models (LLMs) using automated metrics on custom datasets. Learn the best practices for choosing the right LLM for your specific project and see how they perform on various tasks.

👍 Don't Forget to Like, Comment, and Subscribe for More Tutorials!

00:00 - Intro
01:10 - LLM evaluation approaches
05:36 - Available tools & metrics
08:04 - Evaluation process
08:55 - Google Colab setup
09:49 - Dataset
11:25 - Generate model predictions
12:50 - Naive evaluation
14:55 - Use AI to evaluate AI
19:00 - Evaluation report
21:14 - Conclusion

Join this channel to get access to the perks and support my work:

#rag #llama3 #llm #langchain #python #artificialintelligence
Рекомендации по теме
Комментарии
Автор

Excellent explanation and presentation! Well done Sr!

Cyberspider
Автор

Perfect. Content like this reminds me subscribing to your channel and website was a good decision. I'm curious to dig into the prompt they used for their metrics. I checked their documentation, and it allows defining custom metrics. That is something really useful, I will share your video in my X (twitter) account, thx.

unclecode