Mistral AI API - Mixtral 8x7B and Mistral Medium | Tests and First Impression

preview_player
Показать описание
Mistral AI API - Mixtral 8x7B and Mistral Medium | Tests and First Impression

👊 Become a member and get access to GitHub:

Get a FREE 45+ ChatGPT Prompts PDF here:
📧 Join the newsletter:

🌐 My website:

I test the new Mistral AI API vs ChatGPT (OpenAI API), this API includes the Mixtral8x7B MoE model plus and secret model i did not find any info about. Cool testing in Python of the Mistal AI API. Great results!

00:00 Mistral AI API Intro
00:17 Mistral AI Platform + Pricing
02:15 Mistral API vs OpenAI Python Testing
11:22 Mistral API Steaming
12:51 My Conclusion
Рекомендации по теме
Комментарии
Автор

The answer from mixtral medium "ball is lost during transit" doesn't necessarily mean that the ball is lost in the box during shipping, it could be lost during transit between the moment you put the ball in the bag and the moment you put the bag in the box. imho, the model got it right, the human just didn't interpret the result correctly. And the gpt4 answer you labeled as 'perfect' could be wrong as well. Depending on how the bag was tilted, the ball wouldn't have fallen out of the bag. I feel like the mistral-medium answer was the most accurate one.

jorinator
Автор

🎯 Key Takeaways for quick navigation:

00:00 🚀 *Overview and Platform Introduction*
- Mistol AI API access for testing compared to GPT-3.5 and GPT-4.
- Introduction to Mistol AI platform features, including models, streaming options, and safe mode.
- Pricing overview and initial impressions of Mistol AI's competitiveness.
01:24 💰 *Pricing Comparison*
- Detailed pricing calculations for Mistol AI's medium and small models.
- Competitive pricing compared to GPT-3.5 Turbo.
- Ready to proceed with testing after pricing analysis.
02:21 🧠 *Testing Scenarios Introduction*
- Explanation of the three testing scenarios: Shirt problem, World model problem, and Python Snake game.
- Description of the reasoning and coding challenges posed to Mistol AI models.
04:21 🤖 *Testing GPT-3.5 on the Shirt Problem*
- Quick test of GPT-3.5 on the Shirt problem.
- GPT-3.5's incorrect response and analysis of the mistake.
- Setting the stage for Mistol AI's response to the same problem.
05:17 👕 *Testing Mistol Small Model on the Shirt Problem*
- Mistol Small Model's correct response to the Shirt problem.
- Highlighting Mistol's ability to understand parallel processing in the problem.
- Confidence in Mistol's capability based on the small model's performance.
06:08 🌍 *Testing Mistol Medium Model on the World Problem*
- Introduction to the World model problem.
- GPT-3.5's incorrect response to the World problem.
- Preparing to test both Mistol Small and Medium models on the same problem.
07:29 🌐 *Testing Mistol Small and Medium Models on the World Problem*
- Mistol Small Model's response and analysis.
- Mistol Medium Model's response and analysis.
- Comparison with GPT-4's accurate response to the World problem.
08:51 🐍 *Testing Python Snake Game - Mistol Small Model*
- Implementing and testing Python Snake Game code using Mistol Small Model.
- Evaluation of the generated code's quality.
- Comparison with Mistol's response to other models.
10:31 🎮 *Testing Python Snake Game - Mistol Medium Model*
- Implementing and testing Python Snake Game code using Mistol Medium Model.
- Evaluation of the generated code's quality and UI.
- Comparison with Mistol Small Model and GPT-4's responses.
11:42 🔄 *Testing Streaming Function - Mistol Tiny, Small, and Medium Models*
- Introduction to streaming functionality on Mistol AI.
- Quick streaming test on Mistol Tiny, Small, and Medium models.
- Comparison of streaming speed among different Mistol models.
13:05 🌐 *Conclusion and Future Plans*
- Positive feedback on Mistol AI's performance and streaming functionality.
- Expressing excitement about exploring other APIs and supporting Mistol's progress.
- Curiosity about Mistol's Medium model and the potential for more benchmarks and information.

Made with HARPA AI

DJPapzin
Автор

Mixtral is amazing. It's the first one I see that gets this question right: "In a totally classical family, a girl named Sally has 3 brothers Alfred, Bernard and Charlie, who each have 2 sisters. How many sisters does Sally have?" and it even explains it well. MistralAI is really trying to address the issues that plague other LLMs and that's great, it will end up with one we can finally trust a little bit more.

levieux
Автор

🎯 Key Takeaways for quick navigation:

00:00 🤖 *Overview of Mistral AI models and pricing*
02:21 👕 *Comparing models on shirt drying word problem*
- Mistral gets it right, ChatGPT gets it wrong
06:22 🏀 *Comparing models on ball in bag world problem*
- GPT-4 reasons perfectly, Mistral models struggle
08:51 🐍 *Comparing models on coding snake game in Python*
- GPT-4 codes full game, Mistral gives partial code
11:27 ⏩ *Demo of Mistral streaming responses *
- All models stream paragraphs quickly
12:50 👍 *Overall positive, ready to explore more APIs*

Made with HARPA AI

prepthenoodles
Автор

Eagerly anticipating Mistral’s debut in our upcoming Taskade Multi-Agent update! 🌈

Taskade
Автор

Great overview. Thanks for showing the code tests.

onoff
Автор

Thank you for testing it - at the moment I use GPT API - but maybe in 2024 I will try some OpenSource models.

micbab-vgmu
Автор

Thanks a bunch for showing comparisons of the different models and how they preform...

SwizZLe
Автор

Any AI Assistant functionality with tools (RAG, code interpreter and function calling type) you came across from any of the models out there?

nazihfattal
Автор

I appreciate you showing actual pricing numbers, could we perhaps normalize showing specific numbers for things like this as well as for size requirements for local versions (when available) would be very helpful going forward.

michaelpiper
Автор

Does it require a lot of resources to run it locally?

mak_kry
Автор

So it doesn't have a friendly interface like Open AI's playground?

TVdosPatriotas
Автор

How do I finetune SOTA models? They're cool, but they don't allow me to make the most of them. Finetuning would solve that, and I'd pay for that, but they don't have such an option, and setting up everything locally manually is too complicated. GPT4, biggest Mistral model - I want to be able to fine-tune them!

dmy_tro
Автор

By the way, the weights of Mixtral 8x7b are released, so you can run it locally with enough ram/vram.

wurstelei
Автор

Mistral has my sympathy bonus. I am able to run this offline as well.

v-for-victory
Автор

What is the reason programs like this and stable diffusion are not ‘plug and play’ with more simple installation?

lenderzconstable
Автор

is mixtral running locally, is it using the internet at all?

yuske
Автор

Just me, patiently waiting until a model this good is uncensored.

clearandsweet
Автор

Sorry but for the ball in the bag with a hole problem, I'm going to have to give a fail to
GPT-3
GPT-4
Mistral-small
Mistral-medium
GPT-All-About-AI

All of you failed the test as none of the models understood all important aspects of the problem. Out of all models it seems Mistral models had the closest answers, but still missed the small probability that the ball rolled over the hold in the bottom of the bag, dropped out and the person did not notice. Mistral likely just assumed this was too stupid a scenario to consider.

nope
Автор

NGL, around 6-11 months ago I really liked your content, and I thought a channel labeled "All about AI" would also cover any big updates in AI, critically testing and comparing models, and so on. Yet I miss any and all content on new stuff such as Grok, Bard, Gemini, and so on. I just no longer find the content interesting as it never covers the type of AI stuff I am interested in, which is weird considering the channel name. I personally am unsubscribing, but thought before doing so I should state why. best of luck with your channel though, nothing against yourself, you seem like a cool dude.

orbedus